The vacancy is well-defined but lacks compensation details, affecting overall attractiveness.
no salary info
Job description
Beeline is a leading telecommunications company focused on providing innovative solutions and services to its customers.
Responsibilities
### Responsibilities
- Design, develop, and support ETL/ELT pipelines.
- Integrate with various data sources (APIs, databases, files, queues).
- Optimize big data processing (10TB+).
- Work with distributed storage and computing systems (Hadoop, Spark).
- Set up orchestration and monitoring of pipelines in Airflow.
- Support and develop storage (ClickHouse, HDFS, Postgresql).
- Prepare data marts for analysts and ML teams.
- Document pipelines and best practices within the team.
- Develop tools for integrating LLM into products (APIs, libraries, etc.).
- Develop agents and agent systems.
- Create knowledge bases for RAG.
- Fine-tune/adapt LLM for specific tasks/domains (PeFT, SFT).
Requirements
### Requirements
- 2+ years of experience as a Data Engineer.
- Commercial development experience in Python (3.6+).
- Experience with PySpark and the Hadoop ecosystem (HDFS, Yarn).
- Understanding of distributed computing principles and Spark job optimization.
- Practical experience with Airflow (DAGs, sensors, operators, optimization).
- Knowledge of Machine Learning / Deep Learning fundamentals.
- Deep understanding of Transformer architecture.
- Ability to work with large volumes of data (10TB+).
- Knowledge of SQL (including optimization of complex queries).
- Experience with version control systems (we use Git + Gitlab).
About Beeline
Beeline is a telecommunications company in Russia offering job opportunities, including entry-level positions without experience requirements, training programs, predictable income levels, and career development paths through Beeline University. They recruit across various directions and provide hundreds of courses for employee development.