Power through data with Excel and mine for impactful stories that only spreadsheets can disclose with a few clicks. This session will teach you the basics you need to turn tabular data into headlines and to edit your spreadsheet for user-friendly team work. Because every reporter needs to be able to check facts and figures.
For this session you need to bring your computer with a locally installed version of Microsoft Excel. If you dont have that, go to the Data-Pub and get it installed before you enter the session.
This session is about extracting data from webpages without using any code. When data is stuck within webpages, you can't analyse or visualise it. By taking this session, you will learn how to extract data from thousands of pages and from pages that require clicks to see the data. We'll also build a Google sheet with live data inside it.
Power through data with Excel and mine for impactful stories that only spreadsheets can disclose with a few clicks. This session will teach you the basics you need to turn tabular data into headlines and to edit your spreadsheet for user-friendly team work. Because every reporter needs to be able to check facts and figures.
For this session you need to bring your computer with a locally installed version of Microsoft Excel. If you dont have that, go to the Data-Pub and get it installed before you enter the session.
This session is about extracting data from webpages without using any code. Data is stuck within webpages making it unusable. You can't analyse and visualise it when it's stuck in a website. By taking this session, you will learn how to extract this data. Learn how to extract data from thousands of pages at a time, and to grab it from pages that require clicks to see it. We'll also build a Google sheet with live data inside it.
Introduction to databases. Take the next step in data journalism and learn how to work with MS Access. You'll learn how to work with large data sets, to select, sort, and aggregate data. This class works best if you have some knowledge about Excel or similar programs.
For this session you need to bring your computer with a locally installed version of Microsoft Excel and Microsoft Access. If you dont have that, go to the Data-Pub and get it installed before you join us.
Mac does not support Access, so make sure you have a PC.
You can take your journalism many steps further if you learn to code. You can build web scrapers, you can clean and analyze data, and you can build visualizations. If programming doesn't scare you, you'll find that by adding a bit code to many online services, you'll get a much better result.
It's not hard to code. It can actually be quite fun.
In this session you'll get an introduction to the fundamentals of programming. If you plan to go to later sessions on programming web scrapers, and if you have no prior knowledge of programming, you might want to start with this session.
The trainer of the session, Tommy Kaas, is a journalist, who taught himself to code to create web scrapers.
Bring your own laptop to the session.
Here's how to organize your document trail. We see more and more situations in which journalists get lots of electronic documents and need tools for analyzing, sharing, and publishing. DocumentCloud is a good first step into this world. And with just an hour of practical training, you'll be up and running with this free tool.
Please note that this session is based on Access 1.
Introduction to databases. Take the next step in data journalism and learn how to work with MS Access. You'll learn how to work with large data sets, to select, sort, and aggregate data. This class works best if you have some knowledge about Excel or similar programs.
For this session you need to bring your computer with a locally installed version of Microsoft Excel and Microsoft Access. If you dont have that, go to the Data-Pub and get it installed before you join us.
Mac does not support Access, so make sure you have a PC.
Learn how to summarize data in minutes rather than hours on your spreadsheet with pivot tables. This session will be most useful if you're already familiar with the basics of Excel or another spreadsheet program.
Learn about advanced tricks in Excel that will blow your mind and give you control over your data like never before. We'll conquer common data-cleaning issues and much more.
This session will look at basic calculations for overview analysis of data, showing ways to quickly find stories through ratios, rates, percentage change, and summarizing data with pivot tables.
How can you secure your smartphone communications? This session will teach you how to establish encrypted mobile voice and text communication. We cover the following apps: TextSecure, RedPhone and Signal (Whisper). Android and iOS.
Get help with software downloading in the Data Pub Thursday 3.00-4.30 p.m.
You're comfortable doing data analysis in spreadsheets for your stories and managing databases. But now you would like to take your data journalism and analytical skills to the next level. This session introduces using statistics for stories. It will cover a more systematic way of thinking about your approach to data and then applying your new understanding through descriptive statistics with an emphasis on storytelling. Familiarity with spreadsheets is necessary but no experience with statistics is assumed.
Attendees shoud download software before the session: PSPP stats program. It is very similar to SPSS and even runs the SPSS syntax. Also PSPP has a Mac OS X version and a current Windows version.
Both can be found through links here: http://www.gnu.org/software/pspp/get.html.
You can get installation help at the Data Pub Thursday between 3 p.m. and 4.30 p.m.
Open Refine is the best tool to clean really dirty data -- the kind of data in which the same name might be spelled in 30 different ways. It has built-in cleaning tools for analysts and journalists.
An introduction to webscraping with Python: This two-part, hands-on workshop will teach basic newsroom programming concepts using the Python language. We'll cover how to deconstruct a common reporting task -- gathering a table of data from a public website -- and assemble a solution from useful Python libraries that you can use again and again.
Prerequisites: Attendees should be familiar with HTML and the command line and be comfortable with databases and SQL. If you've ever written a string function in Excel ("=left(A2,5)"), you'll be fine.
Python has to be installed at your laptop before the training. Somebody can help you at the “Data Pub” on Thursday.
Get help with software downloading in the Data Pub Thursday 3.00-4.30 p.m.
This session builds on skills learned in Stats for Stories 1. Now that you have a familiar way to thinking statistically about your data, this session introduces the notion of statistical significance and how to use it to back up findings from cross tabulations. In a sense, it helps answer the questions: Is what I'm finding in my data worth reporting?
An introduction to webscraping with Python: This two-part, hands-on workshop will teach basic newsroom programming concepts using the Python language. We'll cover how to deconstruct a common reporting task -- gathering a table of data from a public website -- and assemble a solution from useful Python libraries that you can use again and again.
Prerequisites: Attendees should be familiar with HTML and the command line and be comfortable with databases and SQL. If you've ever written a string function in Excel ("=left(A2,5)"), you'll be fine.
Python has to be installed at your laptop before the training. Somebody can help you at the “Data Pub” on Thursday.
Get help with software downloading in the Data Pub Thursday 3.00-4.30 p.m.
This final stats session, again building on skills learned in Stats for Stories 1 and 2, introduces one of the most powerful statistical procedures in the investigative reporter's toolkit. We'll cover regression analysis and the power it can add to your storytelling.
Here's how to securely communicate with sources and colleagues around the world. This session will teach you how to establish encrypted chat. We will create a DuckDuckGo account, install Pidgin (PC) or Audium (Mac) and the Off The Record (OTR) plugin. Finally, we will exchange fingerprints and talk about general security surrounding the use of encrypted chat. Mac and PC.
In the two CartoDB sessions you'll learn how to create interactive maps for the web with a free tool. CartoDB is an online application, and you won't have to install anything on your computer.
In this first session you'll be introduced to CartoDB. We will geocode addresses and create maps with points. You'll learn the basic map operations. We'll style the map and the info windows and make the map ready for publishing.Get help with software downloading in the Data Pub Thursday 4.30-6.00 p.m.
All data journalists can benefit from learning Structured Query Language or SQL, the language behind database managers. Whether you're hunting for evidence in large databases, needing to mash data tables together, or building an interactive database for your news organization, SQL is a tool for making these data skills possible. This course introduces SQL with an emphasis on how it can help reporters manage data for deeper investigative dives into that data.
These sessions introduce the SQL model and cover the basics of writing SQL queries. Familiarity with spreadsheets is helpful but no experience with SQL or databases is assumed.
Attendees should install software before the session, and they can get installation help at the Data Pub Thursday between 4.30 p.m. and 6 p.m.
Microsoft Access is the way to start and it’s only available on Windows.
If not Access, the Windows users should either install SQL Server Express 2014 found here:https://www.microsoft.com/en-us/download/details.aspx?id=42299.
They should select ExpressAdv 32BIT\SQLEXPRADV_x86_ENU.exe for the correct 32-bit installer or ExpressAdv 64BIT\SQLEXPRADV_x64_ENU.exe for the correct 64-bit installer. If they’re not sure which to use, then have them install the 32-bit by default. The software is free but not open source.
Or they can go open source (and free) by installing PostgreSQL from herehttp://www.postgresql.org/download/. From that page, there are both 32-bit and 64-bit installers. Once PostgreSQL is installed, they should next install Navicat on top of that found here: http://www.navicat.com/download/navicat-for-postgresql. They’ll have to use the 14-day trial version unless they want to purchase it outright.
For Mac users, we’re recommending PostgreSQL for Mac followed by Navicat on top of that. PostreSQL is found here http://www.enterprisedb.com/products-services-training/pgdownload#osx and Navicat here: http://www.navicat.com/download/navicat-for-postgresql.
You can get installation help at the Data Pub Thursday between 3 p.m. and 4.30 p.m.
Here's how to securely communicate with sources and colleagues around the world. This session will teach you how to establish encrypted chat. We will create a DuckDuckGo account, install Pidgin (PC) or Audium (Mac) and the Off The Record (OTR) plugin. Finally, we will exchange fingerprints and talk about general security surrounding the use of encrypted chat. Mac and PC..
In this second of two sessions you'll learn to use more features in CartoDB. You will see how to create a cloropleth map like this, and we will style the polygons so the different colors reflect different values in the dataset, we are visualizing. We will merge datasets inside CartoDB, a technique which will make it easy to upload datasets, and based on those create cloropleth maps. We'll also create visualizations using several maps. To join this session you must attend session 1 or have a basic knowledge of CartoDB.
Bring your own laptop to the session.
Get help with software downloading in the Data Pub Thursday 4.30-6.00 p.m.
This session builds on skills learned in Structured Query Language for Reporters 1. Now that you've written your first SQL queries, expand you analytical power by learning to summarize and aggregate your data through SQL queries. We'll emphasize how SQL commands help you find evidence for your investigations in larger databases.
Get help with Sofware downloading Friday 4.00-5.00 p.m. in the Data Pub.
This class will introduce you to analyzing data for stories by using mapping software. More than just a pretty picture, mapping software can help you uncover stories that others miss.
The sessions will use ArcGIS Online. You will get free licenses so you can use the software when you return to your home country.
This class is divided into two parts. Each session is one-hour long. You are not expected to have any mapping skills beforehand, but if you know how to use a spreadsheet, that would be helpful.
Get help with software downloading in the Data Pub Thursday 4.30-6.00 p.m.
This final SQL session, again building on skills learned in Structure Query Language for Reporters 1 and 2, shows how to write SQL queries that join separate data tables and databases together for more powerful investigative analysis.
Displaying a series of events can be as simple as drawing a straight line. But if you want to get fancier, there are a bunch of other options to display chronologies and storylines. In this workshop, we'll take a tour of current timelines in the wild and walk you through three open-source tools to help you make your own: ProPublica's TimelineSetter, Zach Wise's TimelineJS, and WNYC's Vertical Timeline.
Get help with Sofware downloading Friday 4.00-5.00 p.m. in the Data Pub.
This class will introduce you to analyzing data for stories by using mapping software. More than just a pretty picture, mapping software can help you uncover stories that others miss.
The sessions will use ArcGIS Online. You will get free licenses so you can use the software when you return to your home country.
This class is divided into two parts. Each session is one-hour long. You are not expected to have any mapping skills beforehand, but if you know how to use a spreadsheet, that would be helpful.
An introduction to webscraping with Python: This two-part, hands-on workshop will teach basic newsroom programming concepts using the Python language. We'll cover how to deconstruct a common reporting task -- gathering a table of data from a public website -- and assemble a solution from useful Python libraries that you can use again and again.
Prerequisites: Attendees should be familiar with HTML and the command line and be comfortable with databases and SQL. If you've ever written a string function in Excel ("=left(A2,5)"), you'll be fine.
Python has to be installed at your laptop before the training. Somebody can help you at the “Data Pub” on Thursday.
Displaying a series of events can be as simple as drawing a straight line. But if you want to get fancier, there are a bunch of other options to display chronologies and storylines. In this workshop, we'll take a tour of current timelines in the wild and walk you through three open-source tools to help you make your own: ProPublica's TimelineSetter, Zach Wise's TimelineJS, and WNYC's Vertical Timeline.
An introduction to webscraping with Python: This two-part, hands-on workshop will teach basic newsroom programming concepts using the Python language. We'll cover how to deconstruct a common reporting task -- gathering a table of data from a public website -- and assemble a solution from useful Python libraries that you can use again and again.
Prerequisites: Attendees should be familiar with HTML and the command line and be comfortable with databases and SQL. If you've ever written a string function in Excel ("=left(A2,5)"), you'll be fine.
Python has to be installed at your laptop before the training. Somebody can help you at the “Data Pub” on Thursday.
Importing PDFs and getting structured data out of the files is a challenge. We'll train you to use CometDocs and structuring in Excel. CometDocs is for now the best tool to extract the kind of strange characters we face when converting from different languages. It is free for all members of IRE.
An introduction to webscraping with Python: This two-part, hands-on workshop will teach basic newsroom programming concepts using the Python language. We'll cover how to deconstruct a common reporting task -- gathering a table of data from a public website -- and assemble a solution from useful Python libraries that you can use again and again.
Prerequisites: Attendees should be familiar with HTML and the command line and be comfortable with databases and SQL. If you've ever written a string function in Excel ("=left(A2,5)"), you'll be fine.
Python has to be installed at your laptop before the training. Somebody can help you at the “Data Pub” on Thursday.
In this practical session you will learn to use the geolocation OSINT tool cree.py, created by Greek software developer Ioannis Kakavas. Learn how to trace people's movements throughout the day through social media.