Loading…
Attending this event?
October 22-23, 2024
Santa Clara, CA
View More Details & Registration

The Sched app allows you to build your schedule but is not a substitute for your event registration. You must be registered for RISC-V Summit to participate in the sessions. If you have not registered but would like to join us, please go to the event registration page to purchase a registration.
Intermediate clear filter
arrow_back View All Dates
Tuesday, October 22
 

1:55pm PDT

Lessons Learned in Using RISC-V for Generative AI and Where We Can Go from Here - Jin Kim, Esperanto Technologies
Tuesday October 22, 2024 1:55pm - 2:13pm PDT
The size of the Foundation models behind the Generative AI revolution have grown at a rate of more than 400x every 2 years, while DRAM memory capacity has been increasing only at 2x every two years, leading to what is commonly called the “memory wall”. Similarly, while the required throughput rate of LLMs making up the Foundation models has been increasing at 10x per year, the increase in computational capability of GPUs has been at a pace of only 10x in 4 years, leading to what is commonly called the “compute wall”. These trends have raised a new set of challenges in how to economically train these models, cost-effectively run them, and manage the tremendous increase in electrical power. The first contribution of this session are lessons learned in leveraging hardware and software developed for traditional AI workloads and how it was extended to support Generative AI. The session’s next main contribution is how we are applying lessons learned from our first-generation technology to our next generation. In this session’s final contribution, we will also discuss how the RISC-V ISA could be extended in ways that would make it more efficient and compelling at running Generative AI.
Speakers
avatar for Jin Kim

Jin Kim

Chief Data Science Officer, Esperanto Technologies
An executive, entrepreneur, and data scientist, Jin’s experience spans enterprise software products and services in AI, big data, and advanced analytics. He has led multinational engineering teams at both established and startup companies, including GraphSQL, Wave Computing, Objectivity... Read More →
Tuesday October 22, 2024 1:55pm - 2:13pm PDT
Grand Ballroom G (Level 1)
  AI / ML

2:15pm PDT

Building Tool Chains for RISC-V AI Accelerators - Jeremy Bennett, Embecosm
Tuesday October 22, 2024 2:15pm - 2:33pm PDT
Our client is developing a massively parallel 64-bit chip for AI inference workloads. To facilitate early software development, we are bringing up an AI tool flow for this chip in a QEMU RISC-V environment. In this talk, we'll share our experience of getting three key AI frameworks working with RISC-V QEMU: Pytorch, Tensorflow and the OpenXLA compiler. Our talk will share our experience addressing two key issues. We will describe the challenges we faced, their solutions and reflect on the lessons learned for future work. The first of these is simply getting the tools to effectively run in an emulated RISC-V environment. These tools are large, fast moving pieces of software with extensive external dependencies. Our second challenge is performance. AI workloads are inherently parallel, and hence run efficiently on vector enabled hardware. However RISC-V vector (RVV) is relatively new, and we experienced difficulty getting the performance we expected out of the tool flow. At the end of this talk, we hope our audience will have a better understanding of the challenges in bringing up an AI tool flow under QEMU. We hope out experience will help them bring up their own AI tool flows.
Speakers
avatar for Jeremy Bennett

Jeremy Bennett

Chief Executive, Embecosm
Bio: Dr Jeremy Bennett is founder and Chief Executive of Embecosm(http://www.embecosm.com), a consultancy implementing open sourcecompilers, chip simulators and AI/ML for major corporations around the world.He is a author of the standard textbook "Introduction to CompilingTechniques... Read More →
Tuesday October 22, 2024 2:15pm - 2:33pm PDT
Grand Ballroom G (Level 1)
  AI / ML

2:15pm PDT

An Adaptive Interrupt Architecture for Extremely Timing-Critical Applications - Jamie Kim, Samsung Electronics
Tuesday October 22, 2024 2:15pm - 2:33pm PDT
I would like to introduce our success story of adopting RISC-V CPU in the embedded domain with the ability to customize the architecture. For this success, we developed a scalable, orthogonal and transparent interrupt architecture which enabled the control of extremely timing-critical tasks. I believe this architecture can be widely adopted across multiple domains with the configurability to adopt to their own requirements.
Speakers
avatar for Jamie Kim

Jamie Kim

Principal Engineer, Samsung Electronics
Jamie Kim received Ph.D. degree on Computer Architecture back in 2015 and has been working in System LSI, Samsung ever since. He led multiple MCU projects using RISC-V, which successfully went to mass production, including the first ever RISC-V based product in Samsung. Currently... Read More →
Tuesday October 22, 2024 2:15pm - 2:33pm PDT
Grand Ballroom H (Level 1)

2:15pm PDT

RISC-V CPU Development Using Olympia Performance Model - Knute Lingaard, MIPS
Tuesday October 22, 2024 2:15pm - 2:53pm PDT
The RISC-V Foundation's Olympia Performance Model is a great tool as the basis for designing a high-performance RISC-V CPU design. This session will provide a high-level overview of the Olympia Performance Model and then provide examples of how to use the model for tradeoff analysis on different RISC-V Out-of-Order superscalar designs.
Speakers
avatar for Knute Lingaard

Knute Lingaard

Sr. Principal Engineer, MIPS
Sr. Principal Engineer skilled in performance/functional modeling, software design, C++, and Python. Lead designer and developer of the open source GitHub project Sparcians (https://github.com/sparcians and co-chair of the RISC-V International Performance Modeling SIG
Tuesday October 22, 2024 2:15pm - 2:53pm PDT
Theater (Level 2)
  ISA and Design Tools

2:35pm PDT

LLM Inference on RISC-V Embedded CPUs - Yueh-Feng Lee, Andes Technology
Tuesday October 22, 2024 2:35pm - 2:53pm PDT
The advancement of large language models (LLMs) has significantly enhanced natural language processing capabilities, enabling complex text understanding and generation tasks. This presentation focuses on optimizing the open-source LLaMA CPP project for the RISC-V P extension. By running the TinyLLaMA 1.1B model on the Andes Voyager development board using a quad-core CPU supporting the RISC-V P extension, performance results show that the model can achieve near real-time response. This work highlights the potential of RISC-V as an efficient platform for deploying advanced AI models in resource-constrained environments, contributing to the growing field of edge computing and embedded AI applications.
Speakers
avatar for Yueh-Feng Lee

Yueh-Feng Lee

Manager, Andes Technology
Yueh-Feng Lee received his Ph.D. degree in computer science from National Chiao Tung University. He previously worked at Mediatek and Industrial Technology Research Institute. His areas of focus include AI compiler and runtime, hypervisor technology, and embedded systems.
Tuesday October 22, 2024 2:35pm - 2:53pm PDT
Grand Ballroom G (Level 1)
  AI / ML

2:35pm PDT

Berberis: Dynamic Binary Translation from RISC-V to X86_64 on Android - Lev Rumyantsev & Jeremiah Griffin, Google
Tuesday October 22, 2024 2:35pm - 2:53pm PDT
Berberis is an open source userspace dynamic binary translator facilitating cross-architecture development and testing of RISC-V Android applications. It translates native riscv64 code inside of an Android APK to x86_64 at runtime, enabling developers to test RISC-V builds of their apps on their workstations when target device hardware is unavailable. This presentation will cover the motivation and benefits of userspace translation versus whole-system emulation, the challenges of translating RISC-V code to x86_64, and use cases and future directions for the project.
Speakers
avatar for Jeremiah Griffin

Jeremiah Griffin

Staff Software Engineer, Google
Jeremiah joined Google in 2022 and has been a technical lead of the Berberis RISC-V-to-x86 dynamic binary translator for Android since 2023. His areas of expertise include systems programming, automated testing, embedded and automotive software, and human-machine interfaces. He has... Read More →
avatar for Lev Rumyantsev

Lev Rumyantsev

Software Engineer, Google
Since 2014 Lev has been working at Google on various projects to enhance user experience with Android applications on Large Screen and x86 devices. His main focus has been on developing a binary-translation layer to run ARM-compiled applications on x86 devices. In 2022 Lev also started... Read More →
Tuesday October 22, 2024 2:35pm - 2:53pm PDT
Grand Ballroom H (Level 1)

3:15pm PDT

Development of the First Open-Source Implementation of the RISC-V Vector Cryptography Extension - Markku-Juhani O. Saarinen, Tampere University
Tuesday October 22, 2024 3:15pm - 3:33pm PDT
Version 1.0.0 of the RISC-V Vector Cryptography extensions specification was ratified in late 2023 and adds high-performance cryptography operations to the comprehensive list of ISA features that RISC-V officially supports. We present the first open-source implementation of the RISC-V Vector Cryptography specification, using the PULP Project's Ara vector processor as a baseline and targeting a 28nm technology node. We present the design/verification opportunities and challenges that were encountered. Furthermore, a detailed review of the implementation and benchmarking results will be included in the presentation.
Speakers
avatar for Markku-Juhani O. Saarinen

Markku-Juhani O. Saarinen

Professor of Practice, Tampere University
Markku-Juhani O. Saarinen is a Professor of Practice (työelämäprofessori) at Tampere University (Finland). A cryptographer by training and with a long international career in security engineering, Markku has co-authored many of the ratified RISC-V cryptography extensions. Currently... Read More →
Tuesday October 22, 2024 3:15pm - 3:33pm PDT
Grand Ballroom H (Level 1)

3:15pm PDT

Enhance the Performance of QEMU RVV Load/Store Implementation - Max Chou, SiFive & Jeremy Bennett, Embecosm
Tuesday October 22, 2024 3:15pm - 3:33pm PDT
QEMU is an emulator that developers can developer and debug their software on it before getting the real RISC-V hardware. We observed that vectorized executables run much slower than non-vectorized ones on QEMU. From benchmarks (e.g. SPEC CPU2k6 h264), we can see that most of the execution time is occupied by RVV load/store instructions. The same observation has been reported in the QEMU community. For example, the glibc memcpy benchmark runs 2x to 60x slower than its scalar equivalent on QEMU. We aim to improve the performance of RVV instructions in QEMU, thereby reducing the execution time required for tasks such as Android bootup. In this talk, we will provide an overview of how we enhanced the performance of QEMU RVV load/store instructions and discuss future work.
Speakers
avatar for Jeremy Bennett

Jeremy Bennett

Chief Executive, Embecosm
Bio: Dr Jeremy Bennett is founder and Chief Executive of Embecosm(http://www.embecosm.com), a consultancy implementing open sourcecompilers, chip simulators and AI/ML for major corporations around the world.He is a author of the standard textbook "Introduction to CompilingTechniques... Read More →
avatar for Max Chou

Max Chou

Engineer, SiFive
Max Chou is a Staff Software - Systems Development Engineer at SiFive. His research interests include binary translation, debugging, optimizations, performance and program analysis tools.
Tuesday October 22, 2024 3:15pm - 3:33pm PDT
Theater (Level 2)
  ISA and Design Tools
 
Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.
Filtered by Date -