Data Analytics
- Learn about Azure Stream Analytics
- Learn how to input data into Azure ASA
- Learn how to create queries and data analysis pipelines with Azure ASA
- Learn how to output the results of the data analysis pipeline to various output formats
- Learn how to work with Azure ASA and Microsoft PowerBI
- Learn how to visualize your data
Hadoop: Fundamentals training course is designed to give you the basic overview of the Hadoop framework. The course covers the fundamental concepts to process and analyze large sets of data stored in HDFS. The course also briefly touches the more advanced concepts like Sqoop and Flume for data ingestion. The more details about these advanced concepts is covered under the course Hadoop: Intermediate.
The Hadoop: Fundamentals course is part of a two course series which covers the essential concepts in getting to know Hadoop and the big-data analytics. With increasing digital trend in the world, the importance of big data and data analytics is going to continue growing in the coming years. This course will enable the candidates to explore opportunities in this growing field of digital science.
This course will teach students about Hadoop architecture, ETL, and MapReduce. This course will teach students about Hadoop architecture, ETL, and MapReduce. This course will teach students about Hadoop architecture, ETL, and MapReduce. This course will teach students about Hadoop architecture, ETL, and MapReduce. This course will teach students about Hadoop architecture, ETL, and MapReduce.
Hadoop is becoming an essential tool in the ever-growing Big-Data architecture. This training is designed to benefit:
- Software developers and architects working in Big-Data organizations
- Business and technical analytics professionals
- Senior IT professionals
- Data management professionals
- Project managers
- Data scientists
Hadoop: Intermediate training course is designed to give you in-depth knowledge about the Hadoop framework discussed in our Hadoop and MapReduce Fundamentals course. The course covers the concepts to process and analyze large sets of data stored in HDFS. It teaches Sqoop and Flume for data ingestion.
The Hadoop: Intermediate course is part of a two course series which covers the essential concepts in getting to know Hadoop and the big-data analytics. With increasing digital trend in the world, the importance of big data and data analytics is going to continue growing in the coming years. This course will enable the candidates to explore opportunities in this growing field of digital science.
Hadoop is becoming an essential tool in the ever-growing Big-Data architecture. This training is designed to benefit:
- Software developers and architects working in Big-Data organizations
- Business and technical analytics professionals
- Senior IT professionals
- Data management professionals
- Project managers
- Data scientists
In this short and concise course you will learn everything to get started with web scraping using Java.
You will learn the concepts behind web scraping that you can apply to practically any web page (static AND dynamic / AJAX).
Course structure
We start with an overview of what web scraping is and what you can do with it.
Then we explain the difference in scraping static pages vs dynamic / AJAX pages. You learn how to classify a website in one of the two categories and then apply the right concept in order to scrape the data you want.
Now you will learn how to export the scraped data either as CSV or JSON. These are some popular formats that can be used for further processing.
Unfortunately many websites try to block scrapers or sometimes you just do not want to be detected. In the section going undercover you will learn how to stay undetected and avoid getting blocked.
At the end of the course you can download the full source code of all the lectures and we discuss an outlook to some advanced topics (private proxies, cloud deployment, multi threading ...). Those advanced topics are covered in a follow up course I am going to teach.
Why you should take this course
Stop imagining you can scrape data from websites and use the skills for your next web project, you can do it now.
- Stay ahead of your competition
- Be more efficient and automate tedious, manual tasks
- Increase your value by adding web scraping to your skill set
- You should already be familiar with Java and Maven at a basic to medium level (the course will not show you how to setup Java, Maven or an IDE)
- You should be familiar with HTML/CSS and know how to use your browser's developer tools
- You should know about CSS selectors, though we use them for scraping static web pages
- Prior knowledge of jQuery helps you getting started faster with Jsoup, though this is not required
- You should know what a web API and AJAX is (basic level is enough)
APIs are available on most modern websites, and provide an easy way to integrate the websites functionalities into your code. The API that will be focused on is the Twitter API, which which will be used to mine tweets about the event.
Social media, especially twitter, is becoming a hot topic among many investors, as its trends can often predict behavior of the stock market. This course will focus on how Twitter data can be live streamed, and will feature a worked example of the Yahoo hack, that was revealed on December 14th, 2016.
- Contact the Twitter REST API
- Get insights on how rich the information in social media is
- Format requests to contact other REST APIs
- Understand a JSON response
- Be able to get information out of a JSON response
- Understand how to get data from social media and integrate it into code
- Apply their knowledge do stream data live from social media and make effective use of it in their code
Scraping data from webpages can be a tedious job. But it doesn’t have to be.
With Scrapy, you can scrape using XPath or CSS. With the large number of examples from both techniques, you’re sure to find a solution that fits for you.
Whether your targeting data on a single page or multiple, Scrapy can handle the job. No matter if the data is within a list, you can scrape specific patterns right out of the list. Building up your specific Scrapy job isn't a difficult task.
Scrapy is a Python library. If you're familiar with Python, XPath or CSS, you'll feel right at home using Scrapy.
At the end of this course, you will understand:
- what Scrapy is used for
- how to install it
- how to use Scapy
In summary, you'll be able to target specific elements on a webpage, whether the element is stand along or in a list. Then you can retrieve a group of those elements or just one. This technique allows you to pull down specific types of data.
The course ends with a project to help solidify what you've learned. There is a full walk through included with the project solution.