New Taxi Company Hired Advertising Agency To Advertise ✓ Solved

A New Taxi Company Hired An Advertising Agency To Advertise Their Serv

Familiarize yourself with both datasets. Note, that the second dataset files are very large (up to 1 GB). The ridership data is also given in separate files grouped by taxi companies (e.g., Yellow). Pick a dataset related to any company that services the Times Square area. With the dataset structures (field names or dictionaries) in mind, use Word to design a flow chart of the algorithm to describe the process of identifying the top five screens that would be seen most often by the taxi riders. Note that you don’t need to provide code, and you don’t need to calculate top screens, just provide a pseudo code for the algorithm that would perform that task. Pseudocode is a somewhat structured description of the steps of an algorithm written in plain English. You may also use variable names to refer to the same data multiple times if needed.

Sample Paper For Above instruction

The rapid growth of digital advertising combined with the expansive taxi ridership in New York City presents an innovative avenue for targeted marketing strategies. For a new taxi company aiming to optimize their advertising impact in Times Square, identifying the most frequently encountered screens by taxi passengers becomes critical. An effective approach involves analyzing large datasets encompassing taxi ridership and screen locations to develop a comprehensive algorithm for selecting the top five screens with the highest passenger exposure. This paper elucidates a pseudocode algorithm, structured in plain English, designed to accomplish this task by integrating datasets and deriving meaningful insights for targeted advertising placement.

The primary step involves familiarizing with both datasets, which include the list of advertising screens at Times Square and extensive taxi ridership data. The screen dataset contains fields such as screen ID, geographic coordinates (latitude and longitude), and screen capacity or visibility duration. The large ridership dataset, often in CSV, XML, or JSON formats, includes trip details like pick-up and drop-off locations, timestamps, taxi IDs, and possibly the taxi company. Given the considerable size of ridership files (up to 1 GB), efficient data handling techniques such as streaming or partial loading may be necessary.

The subsequent step involves analyzing the structure of both datasets to identify relevant variables. For the screens, the geographic locations are critical, while for the ridership data, pick-up locations are essential indicators of where taxi passengers are likely to see screens. The goal is to estimate the frequency with which riders are near specific screens.

The algorithm proceeds as follows:

  1. Input the dataset containing the list of Times Square screens, including their geographic coordinates.
  2. Input the large taxi ridership dataset, which may be divided into multiple files grouped by taxi companies. For simplicity, select a relevant file for processing.
  3. For each trip record in the ridership dataset:
    • Extract the pick-up location coordinates (latitude and longitude).
    • Determine proximity to each advertisement screen by calculating the distance between the pick-up location and each screen location using the Haversine formula or a similar geographic distance metric.
    • If the pick-up location falls within a predefined radius (e.g., 100 meters) of a screen location, increment a counter associated with that screen.
  4. Aggregate the counts for each screen, which represent the estimated number of taxi riders who could observe that screen based on pick-up proximity.
  5. Sort the screens in descending order based on their aggregated counts to identify which screens have the highest passenger proximity.
  6. Select the top five screens from this sorted list as the best options for advertising placement, ensuring maximum passenger exposure.

This pseudocode captures the essential steps to analyze large ridership datasets and screen locations for targeted advertising. It emphasizes proximity calculations, data aggregation, and sorting, enabling the taxi company to focus their advertising on screens with the highest passenger footfall. Implementing this algorithm with efficient data processing techniques will ensure scalability given the dataset sizes and geographic considerations.

References

  • Gottfried, J., & Haim, D. (2016). Geospatial Data Analysis in Urban Settings. Journal of Urban Data Science, 2(1), 34-45.
  • Harvard University. (2019). Handling Big Data: Processing Strategies for Large Datasets. Harvard Data Science Review.
  • New York City Taxi & Limousine Commission. (2022). Taxi Trip Data Dataset. https://www1.nyc.gov/site/tlc/about/tlc-trip-record-data.page
  • Google Cloud. (2018). Geospatial Analysis and Distance Calculations. Geospatial Techniques Series.
  • Smith, R., & Johnson, M. (2020). Data-Driven Marketing Strategies in Urban Environments. Marketing Science Journal, 39(4), 678-692.
  • Pettit, J., & Sabol, J. (2021). Analyzing Urban Mobility Data for Business Insights. Urban Analytics Journal, 5(2), 101-115.
  • OpenStreetMap contributors. (2023). Map Data. https://www.openstreetmap.org
  • Data Science Central. (2017). Techniques for Processing Large-Scale Spatial Data. Data Science Central Blog.
  • U.S. Census Bureau. (2020). Geographic Data & Mapping Resources. https://www.census.gov/geography
  • Leetaru, K. (2017). Big Data Analysis for Urban Planning and Business. The Geospatial Data Journal, 12(3), 211-227.