Descripción del puesto
KEY RESPONSIBILITIES
- Responsible for extracting processing grouping and analyzing data generating the relevant reports.
- Create automated processes where possible to complete certain tasks involving large amounts of data.
- Communicate with internal customers to resolve inconsistencies found.
- Identify opportunities for improvement by increasing efficiency.
- Detailed review of files and import data to the production database.
- Maintain and ensure the integrity of the internal bases.
- Elaboration of Dashboards from different data sources.
- Database & Query Optimization: Design tune and optimize complex SQL queries to improve performance and reduce execution times on production databases.
- Advanced Data Cleaning: work on deep debugging processes data transformation and strict validation of metrics and formats in large volumes of data using Python/SQL.
- API Data Consumption: Extract process and analyze data retrieved from third-party REST APIs to incorporate into reporting and analytical workflows.
- High-Performance Processing: Design and implement highly efficient data pipelines utilizing vectorized operations to minimize execution time on large datasets.
- Concurrent Workflows: Optimize data extraction and API consumption by implementing asynchronous programming and multithreading/multiprocessing solutions.
- Server & Environment Management: Deploy scripts manage remote connections via CLI (SSH) and configure Python virtual environments.
Requisitos
PREVIOUS EXPERIENCE REQUIRED
- 2 years of previous experience in similar jobs.
- Experience with SQL (MySQL PostgreSQL) including proven experience in Performance Tuning and database optimization.
- Experience with Python (Pandas Numpy Plotly Requests logger etc.) with a strong focus on vectorized operations and dependency management using virtual environments (venv conda).
- Experience with concurrent execution (AsyncIO Multithreading) to accelerate data workflows.
- Practical experience consuming REST APIs.
- Practical experience with Command Line Interfaces (CLI) including SSH access checking active connections and basic server navigation.
- Experience with Notebooks (Jupyter Collab).
- Big data analysis.
KNOWLEDGE
- Microsoft 365 tools at an advanced level especially Excel (pivot tables).
- Extra data analysis tools (R Tableau Power BI & Salesforce) will be considered a plus.
- Computer skills and ability to quickly learn new technologies.
- FLUENT advanced conversational English is REQUIRED.
Detalles
Tags:
dataanalyst pandas numpy plotly mysql r tableau powerbi bigdata
Required Experience:
IC
Descripción del puesto KEY RESPONSIBILITIESResponsible for extracting processing grouping and analyzing data generating the relevant reports.Create automated processes where possible to complete certain tasks involving large amounts of data.Communic...
Descripción del puesto
KEY RESPONSIBILITIES
- Responsible for extracting processing grouping and analyzing data generating the relevant reports.
- Create automated processes where possible to complete certain tasks involving large amounts of data.
- Communicate with internal customers to resolve inconsistencies found.
- Identify opportunities for improvement by increasing efficiency.
- Detailed review of files and import data to the production database.
- Maintain and ensure the integrity of the internal bases.
- Elaboration of Dashboards from different data sources.
- Database & Query Optimization: Design tune and optimize complex SQL queries to improve performance and reduce execution times on production databases.
- Advanced Data Cleaning: work on deep debugging processes data transformation and strict validation of metrics and formats in large volumes of data using Python/SQL.
- API Data Consumption: Extract process and analyze data retrieved from third-party REST APIs to incorporate into reporting and analytical workflows.
- High-Performance Processing: Design and implement highly efficient data pipelines utilizing vectorized operations to minimize execution time on large datasets.
- Concurrent Workflows: Optimize data extraction and API consumption by implementing asynchronous programming and multithreading/multiprocessing solutions.
- Server & Environment Management: Deploy scripts manage remote connections via CLI (SSH) and configure Python virtual environments.
Requisitos
PREVIOUS EXPERIENCE REQUIRED
- 2 years of previous experience in similar jobs.
- Experience with SQL (MySQL PostgreSQL) including proven experience in Performance Tuning and database optimization.
- Experience with Python (Pandas Numpy Plotly Requests logger etc.) with a strong focus on vectorized operations and dependency management using virtual environments (venv conda).
- Experience with concurrent execution (AsyncIO Multithreading) to accelerate data workflows.
- Practical experience consuming REST APIs.
- Practical experience with Command Line Interfaces (CLI) including SSH access checking active connections and basic server navigation.
- Experience with Notebooks (Jupyter Collab).
- Big data analysis.
KNOWLEDGE
- Microsoft 365 tools at an advanced level especially Excel (pivot tables).
- Extra data analysis tools (R Tableau Power BI & Salesforce) will be considered a plus.
- Computer skills and ability to quickly learn new technologies.
- FLUENT advanced conversational English is REQUIRED.
Detalles
Tags:
dataanalyst pandas numpy plotly mysql r tableau powerbi bigdata
Required Experience:
IC
View more
View less