When is information too much information? Well, if you work with data, probably never. But, clients and research report writers often need summarised information ready and waiting for KPI and ROI purposes. In our previous post, we created this informative report of which quote contained the most words in ascending order. All well and good.Continue reading “Perl to please the client mind”
Why you pushing me?
Delete your duplicates Welcome to part 4. In the previous posts we looked at how to inspect the validity of a CSV file, how to determine if it fits nicely in a spreadsheet, check whether it will create more columns than needed. Also, we wrote some simple conditions to check year and rating. We thenContinue reading “Adventures in Perl part 4”
Data duplication or not duplication… that is the question In parts 1 and 2 we looked at messy CSV data and how to identify it with Perl. We’ve been using a small CSV file so we can actually see its entire contents; the good, the bad, and the ugly lines – and how Perl canContinue reading “Adventures in Perl part 3”
The CSV file – Trust No One That’s right. Do as Fox Mulder from The X-Files would do. When you are handed a CSV file, don’t trust it. I once merrily loaded sports data into a system taking it at face value, only to discover that an individual whose swim lap times could rival MichaelContinue reading “Adventures in Perl part 2”
The CSV file – quotes and terminators – does my spreadsheet look good? There’s no getting away from the CSV file; readable in text editors and spreadsheets on Mac, Windows and Linux. Spreadsheets being the most popular method to read CSV files is hardly surprising. Open up Excel or LibreOffice Calc and we get tidy,Continue reading “Adventures in Perl part 1”