I am a freelance Senior Perl Developer with over 20 years' experience, including exposure to Machine Learning, Python, Git, MySQL, Nginx, Apache HTTP Server, XSLT, XML, RelaxNG, HTML, and CSS.
I'm an active proponent of Modern Perl. I like writing technical documentation and unit tests; both have saved my customers and me a lot of time over the years. I prefer to reuse tested code as much as possible, hence I often start a project with researching available solutions on CPAN.
My personal development projects consist of several in-house tools, including a static website and blog generator using XML as input format. Currently, I am working on a new version using Markdown for input instead.
Over the years many projects I've been working on involved crawling, caching, parsing, and storing the desired result in a database. I like the challenges such projects offer, like making the web crawlers robust and be able to deal with the intricacies of the Internet and its data. Regarding the latter, this is why I prefer strict parsing of data, with many checks and extensive logging using Log::Log4perl and good test coverage. This way changes to the HTML of scraped web pages, which often occurs in my experience, are detected early and collecting invalid data or missing additional data might be avoided.
Besides parsing HTML I’ve also experience with parsing XML, several custom formats, and domain-specific languages (DSLs).
Another task that often is assigned to me is the conversion of data from one format to another; including data cleaning and verification. For example, modifying the text output of a legacy application in such a way that the new output is suitable to be printed on redesigned labels. Often the input format has to be reverse engineered and documented; a challenge I like.
I prefer to use extensive testing of the custom modules and scripts I write. My favorite module is Test::Most because the module provides the most commonly used testing functions with a single line, avoiding a lot of boilerplate.
My main Perl development environment is currently Ubuntu 17.04 Zesty Zapus running in a virtual machine on OS X. My router runs OpenWrt, and I use a fanless computer to study FreeBSD. I also use two different VPS providers for hosting my websites. I do the administration of each of those systems, e.g. updates, firewall rules, and other security related tasks, installing and configuring software e.g. Apache HTTP server, Nginx, Postfix, Dovecot, OpenSSH, and MySQL.
I like to write technical documentation. I prefer to use markup languages like Org Mode (Emacs), Markdown, and LaTeX. The use of plain text makes version control and searching from the command line easily. Plain text also makes it very easy to generate code from the documentation and vice versa. Where required, PDF versions can be generated using tools like Emacs, Pandoc, and pdflatex.
I exclusively work remotely. Over the past years, I have worked with customers in Japan, USA, The Netherlands, and Canada. Working from home provides me with a productive environment with a minimal number of interruptions. I use the Internet to stay in touch with peers and my craft. I prefer to communicate using email. I have experience with encrypted email using GnuPG, SSH with public key encryption, Git, Subversion, and Github.
I would be happy to discuss my experience with and exposure to: Python, MySQL, XML, XSLT, XSL-FO, RelaxNG, Apache HTTP Server, Nginx, HTML, CSS, Search Engine Optimization (SEO).
Senior Perl Developer; August 2015 - Present; remote, part-time
Modern Perl, LWP::UserAgent, Log::Log4perl, Test::Most, Text::Diff, XML::Parser, DBI, Config::Tiny, Try::Tiny, Path::Tiny, Plack, JSON, git, MySQL, GitHub, Emacs Org Mode, Markdown, Nginx, SEO.
Developed a program to import a "JSON per line" database dump into a MySQL database. I reverse engineered the file format, designed the 70+ MySQL database tables, and added logic to avoid duplicate code.
Developed a web crawler that downloads XML data, verifies this data, gathers some statistics, and stores desired data and statistics in a MySQL database. I am also reponsible for writing all the technical documentation, specifications, and example SQL queries. (ongoing)
Developed several tools for analyzing search engine result pages (SERPs) (ongoing)
Because the web crawler program downloads data from the Internet it must be able to handle timeouts, partial downloads, etc. Moreover the XML parser must be able to distinguish between errors that can be safely ignored because they won’t affect the desired data much, and fatal errors. As a single crawl can gather a lot of data, hence the MySQL database must be designed to deal with querying tens of millions of rows of data in mind.
Senior Perl Developer; February 2013 - December 2015; remote, part-time
Modern Perl, LWP::UserAgent, Log::Log4perl, HTML::TreeBuilder, Test::Most, Test::More, Test::Output, JSON::XS, DBI, Try::Tiny, XML::Parser, XML::Writer, Email::Sender, Text::CSV_XS, MaxMind::DB::Reader, Dancer2, MySQL, RelaxNG, XSLT, Apache Ant, git, VirtualBox, nginx, cron, Emacs Org Mode, OpenSSL, iptables, HTML, CSS, GNU Privacy Guard (GPG).
Developed several web crawling Perl programs for gathering security related data, each crawler exporting downloaded data as a CSV file.
Developed a Perl program that downloads security related data from several sources, verifies, cleans, normalizes the data, and export it as an XML file.
Developed a Perl program that post-processes the exported XML data to remove duplicate entries and overlapping IP ranges using a binary search algorithm.
Developed a Perl program that parses log files (Log4Perl) generated by aforementioned programs, creates reports, and emails those reports.
Developed a Dancer2 web application to manually enter IP ranges and related security data and store this information into a MySQL database.
Configured virtual machines for each of the Perl programs, including designing firewall rules, and allowing access via rsync and HTTPs in a secure manner where required. Documented the entire process of configuration and installation in great detail.
Wrote documentation on how to create certificates to allow client side certificate authentication in nginx (HTTPS), and configuring nginx accordingly.
Wrote extensive tests for all developed custom Perl modules and scripts.
Designed file formats (XML, CSV). Documented each format in Org Mode format (Emacs).
Designed Relax NG schemata to validate XML files.
Wrote XSLT to transform XML output to older formats in order to support client programs that require an older format prior to being safely updated.
Wrote extensive technical documentation in Org Mode format (Emacs).
Because the generated XML file is used in security related applications great care had to be taken to verify the integrity of the downloaded data and make sure that malformed data wouldn’t end up in the final XML file. Because the generated data is used in programs written and maintained by other programmers, clear specifications and well-written documentation was mandatory. In order to keep this documentation under version control Emacs’ Org Mode was used to write in plain text and convert it to, for example, PDF. Due to the asynchonous nature of the various programs care had to be taken to avoid race conditions.
Senior Perl Developer; November 2011 - Present; remote, part-time
Modern Perl, CGI, DBI, Image::Magic, Test::Most, Util::Image, Apache HTTP server, HTML, CSS. MySQL, git, subversion.
Maintenance of several legacy Perl CGI programs. I created new functionality like scaling of photos and adding watermarks to photos. I did various MySQL database optimizations which greatly improved the search query time. I also added new search options.
The Perl CGI program is mostly legacy code which makes extending and/or modifying it quite a challenge. Custom Perl modules are used to group both new functionality and factored out code logically and to support testing.
Computer Science, BSc; 1991.
Design of Algorithms, MSc; 1994. Thesis: "Constraint Satisfaction with Learning Automata."
LaTeX resume template for Pandoc based on a LaTeX resume by Jason R. Blevins. The template can be used to create either a LaTeX or PDF file given a Markdown file as input. The template was used to create this resume. Repository: resume-pandoc.
This repository currently contains a Perl module, Amazon::List, and a small Perl program which uses this module to obtain information from items on an Amazon list and print those formatted to
stdout. Repository: amazon-list
Technical Editor; ISBN-13: 978-0470414644; 2009.
I did some voluntarily reviewing, and am listed in the Credits section. Electronic versions.
I provided the Perl program used in this study; article in Psychosomatics 55(6); June 2014; abstract.
I assisted with Perl programming; article in International Research Journal of Microbiology (IRJM); ISSN: 2141-5463; Vol. 3(11) pp. 363-372; November 2012; full length research paper (PDF)
Dutch (native), English (fluent).