
Julius Smith
Professor Emeritus
Stanford University
About Me
Julius O. Smith is a research engineer, educator, and musician devoted primarily to developing new technologies for music and audio signal processing. He received the B.S.E.E. degree from Rice University in 1975 (Control, Circuits, and Communication), and the M.S. and Ph.D. degrees in E.E. from Stanford University, in 1978 and 1983, respectively. For his MS/EE, he focused largely on statistical signal processing. His Ph.D. research was devoted to improved methods for digital filter design and system identification applied to music and audio systems, particularly the violin. From 1975 to 1977 he worked in the Signal Processing Department at ESL, Sunnyvale, CA, on systems for digital communications. From 1982 to 1986 he was with the Adaptive Systems Department at Systems Control Technology, Palo Alto, CA, where he worked in the areas of adaptive filtering and spectral estimation. From 1986 to 1991 he was employed at NeXT Computer, Inc., responsible for sound, music, and signal processing software for the NeXT computer workstation. After NeXT, he became a Professor at the Center for Computer Research in Music and Acoustics (CCRMA) at Stanford, with a courtesy appointment in EE, teaching courses and pursuing/supervising research related to signal processing techniques applied to music and audio systems. At varying part-time levels, he was a founding consultant for Staccato Systems, Shazam Inc., and moForte Inc. He is presently a Professor Emeritus of Music and by courtesy Electrical Engineering at Stanford, and a perennial consultant for moForte Inc. and a few others. For more information, see https//ccrma.stanford.edu/~jos/.
Sessions
-
Python Templates for Neural Image Classification and Spectral Audio Processing
Lightning Hydra Template Extended and Neural Spectral Modeling Template16:00 - 16:40 UTC | Friday 26th September 2025 | ADCx GatherIntermediateThis presentation introduces two open-source research frameworks for neural image classification and spectral audio processing: (1) the Lightning Hydra Template Extended (LHTE) and (2) the Neural Spectral Modeling Template (NSMT). The LHTE extends the widely used PyTorch Lightning + Hydra template with state-of-the-art architectures (CNNs, ConvNeXt, EfficientNet, Vision Transformers) and expanded dataset support, adding CIFAR-10, CIFAR-100, and a new generalized Variable Image Multi-Head (VIMH) format. VIMH accommodates extremely large image/channel dimensions, multi-head tasks, and supports both classification and regression from a single shared backbone. The LHTE also provides reproducible benchmark experiments, and systematic workflows for rapid model comparison. Built upon […]
-
scipy.cpp
Using AI to Port Python's scipy.signal Filter-Related Functions to C++ for Use in Real Time20:30 - 20:50 UTC | Friday 1st November 2024 | ADCx GatherOnline OnlyThis is a progress report on the following evolving chatbot workflow: "Translate this Python function and its helpers to C++ .h and .cpp files, converting any doc strings to C++ Doxygen comments: ..." [Claude 3.5 Sonnet is hard to beat] "Generate a progressive sequence of unit tests in Catch2 format" [ChatGPT-o1 can be amazing] In general, Python translates smoothly to C++. The chatbots are especially strong in knowing API details, command-line options, and modern C++ idioms. The biggest pitfall seems to be complex algebraic manipulations. When one-shot inference fails (ChatGPT-4o, Claude 3.5 Sonnet), o1 should be tried.