סקירה כללית

^^משרה זו נלקחה מ Career^^NextSilicon is reimagining high
• performance computing (HPC & AI). Our accelerated compute solutions leverage intelligent adaptive algorithms to vastly accelerate supercomputers, driving them forward into a new generation. We have developed a novel software
• defined hardware architecture that is achieving significant advancements in both the HPC and AI domains. At NextSilicon, everything we do is guided by three core values: Professionalism: We strive for exceptional results through professionalism and unwavering dedication to quality and performance. Unity: Collaboration is key to success. That's why we foster a work environment where every employee can feel valued and heard. Impact: We're passionate about developing technologies that make a meaningful impact on industries, communities, and individuals worldwide. impact on industries, communities, and individuals worldwide. We are seeking a highly skilled AI Application Team Lead to build and lead a team responsible for developing, running, and optimizing large
• scale AI workloads on NextSilicon’s AI hardware platform. This role focuses on benchmarking state
• of
• the
• art models (e.g., LLaMA, DeepSeek), executing MLPerf suites, analyzing system
• level performance, and driving cross
• stack optimizations across hardware, runtime, and software frameworks. The ideal candidate combines strong technical depth in AI/ML systems, hands
• on experience with LLM workloads, and leadership capability to guide a high
• performance engineering team. Requirements 5+ years of experience in AI/ML engineering, performance optimization, or ML systems. Deep understanding of LLM architectures, training & inference mechanics, and modern ML frameworks. Strong proficiency in PyTorch ecosystem, with a specific focus on performance tuning via Triton, Cuda or MLIR
• based compiler frameworks. Hands
• on expertise profiling and optimizing kernels (GEMM, attention, softmax, token pipelines). Demonstrated experience running or tuning MLPerf or similar large
• scale benchmarks. Strong Python and C++ development skills. Proven leadership experience: mentoring, guiding, or managing engineers. Responsibilities Lead and mentor a team of AI application and performance engineers. Run and optimize AI workloads (LLaMA, DeepSeek, etc.) and execute MLPerf benchmarks. Analyze end
• to
• end performance and identify HW/SW bottlenecks. Develop optimization strategies across models, kernels, frameworks, and runtime. Build profiling, debugging, and validation tools for large
• scale AI workloads. Collaborate with hardware, compiler, and device software teams to improve performance.

דרישות המשרה

5+ years of experience in AI/ML engineering, performance optimization, or ML systems. Deep understanding of LLM architectures, training & inference mechanics, and modern ML frameworks. Strong proficiency in PyTorch ecosystem, with a specific focus on performance tuning via Triton, Cuda or MLIR
• based compiler frameworks. Hands
• on expertise profiling and optimizing kernels (GEMM, attention, soft