Design and develop data platform architecture, ensuring the stability and high availability of data platform services.
Design, develop, and continuously optimize functionalities of big data components such as offline/real-time data processing, data governance, self-service analytics reports, and OLAP query engines.
Explore the data agent system for large models, building Agents with natural language understanding and data operation capabilities to enhance the intelligence of data analysis, processing, and pipeline construction.
Establish a full lifecycle management system for training data, covering metadata, lineage, quality assessment, and automated evaluation, ensuring controllable and high-quality data for large model training.
Job Requirements
Over 5 years of experience in big data platform-related work, with a bachelor&aposs degree or above in computer science, software engineering, artificial intelligence, mathematics, or related fields.
Proficient in mainstream frameworks such as Spring Boot and MyBatis, with good web development skills (Vue); experience in enterprise-level big data system software integration and development is preferred.
Familiar with the features and application scenarios of components such as Hadoop, Spark, Flink, Kafka, Paimon, ClickHouse, Doris, and able to solve problems encountered during use.
Experience in building large-model data systems, multi-modal data processing systems, and interest in LLM Agents, AutoML, and intelligent data processing workflows, with the ability to explore and implement solutions preferred.
Familiar with large model applications and related technologies such as RAG, MCP, Agent, etc.
Highly responsible, quick learner, good communicator, thoughtful, with strong teamwork skills; team management experience is a plus.