Job Description
Wubble.ai is a pioneering Music AI platform redefining the future of music generation. Our proprietary foundation model pushes the boundaries of AI-driven music creation. We’re proud to work with top-tier global clients, including Disney, Starbucks, Microsoft, HP, and more. Backed by industry giants like Antler, Google, NVIDIA, and others, we are ranked as one of the top 5 startups in Asia!
Role Overview
We are looking for a Contract MLOps Expert with deep experience in CUDA kernel optimization to enable concurrent inference requests on our advanced music generation model. This is a high-impact role where you will directly influence the scalability and performance of our platform.
Key Responsibilities
- CUDA Kernel Optimization : Analyze the existing CUDA kernel implementation and introduce concurrency optimizations to handle multiple requests efficiently.
- High-Performance Computing (HPC): Utilize parallel computing techniques and GPU best practices to maximize throughput and minimize latency.
- MLOps & Integration : Collaborate with engineering teams to integrate optimized kernels into our production environment and CI/CD pipelines.
- Performance Benchmarking : Design and run performance tests to measure the improvements from your optimizations.
- Documentation & Knowledge Transfer: Provide clear technical documentation and best practices for ongoing maintainability.
Qualifications
- GPU & CUDA Expertise : Demonstrated experience in writing, debugging, and optimizing CUDA kernels at scale. Familiarity with concurrency, memory management, and parallelization strategies for GPU-accelerated applications.
- MLOps & ML Frameworks : Hands-on experience with MLOps workflows, CI/CD, and containerization (Docker, Kubernetes). . Proficiency in at least one deep learning framework (e.g., PyTorch).
- High-Scale Model Inference : Experience optimizing large-scale model inference pipelines for real-time or near real-time use cases.
- Problem-Solving & Communication
- Strong analytical skills for troubleshooting performance bottlenecks.
- Ability to communicate technical solutions clearly to both technical and non-technical stakeholders.
- Experience with H100 or similar high-end GPU architectures.
- Familiarity with large language or generative models in the music or media domain.
Why Join Wubble.ai?
- Elite Client Portfolio: Work on solutions that power the creative experiences of world-renowned brands like Disney, Starbucks, Microsoft, and HP.
- Top-Tier Backing: We’re supported by industry leaders such as Antler, Google, and NVIDIA, offering you the opportunity to collaborate with a well-funded and visionary team.
- Cutting-Edge Tech: Contribute to an advanced foundational model pushing the envelope in AI-driven music generation.
- High Impact, High Reward: Your expertise will directly shape the performance and scalability of a groundbreaking platform—this is not your everyday startup gig.
- Remote Collaboration: Enjoy the flexibility of a fully remote contract, enabling you to collaborate from anywhere while tackling exciting, high-profile challenges.
Contract Details
- Contract Type: Contract / Consultancy
- Duration: 1 month, with potential extension based on project needs
- Compensation: Competitive rate, commensurate with experience
How to Apply
Send your resume, portfolio, and a brief cover letter highlighting relevant CUDA optimization and MLOps experience to sufi@wubble.ai with the subject line: "MLOps Expert – CUDA Kernel Optimization Application"
Join Wubble.ai and help us orchestrate the future of music through unparalleled AI innovation!
Similar Jobs
Explore other opportunities that match your interests
last mile
adobe photoshop - tips & trick...