Invited Presentations
- Keynote Talk: Ten Years of ISMIR: Reflections on Challenges and Opportunities
- Keynote Talk: Wind Instrument Playing Humanoid Robots
- Panel Discussion: Industrial Panel Discussion
Keynote Talk: Ten Years of ISMIR: Reflections on Challenges and Opportunities
09:00-10:00 Tuesday (October 27)
Speakers:
- J. Stephen Downie (University of Illinois at Urbana-Champaign, USA)
- Donald Byrd (Indiana University at Bloomington, USA)
- Tim Crawford (Goldsmiths College, University of London, UK)
Abstract:
The International Symposium on Music Information Retrieval (ISMIR) was born on 13 August 1999.
This invited paper expresses the opinions of three of ISMIR's founders as they reflect upon what has happened during its first decade.
The paper provides the background context for the events that led to the establishment of ISMIR.
We highlight the first ISMIR, held in Plymouth, MA in October of 2000,
and use it to elucidate key trends that have influenced subsequent ISMIRs.
Indicators of growth and success drawn from ISMIR publication data are pre-sented.
The role that the Music Information Retrieval Evaluation eXchange (MIREX) has played at ISMIR is also examined.
The factors contributing to ISMIR's growth and success are enumerated.
The paper concludes with a set of challenges and opportunities that the newly formed
International Society for Music Information Retrieval should embrace
to ensure the future vitality of the conference series and the ISMIR community.
Biographies
J. Stephen Downie is an Associate Professor at the Graduate School of Library and Information Science,
University of Illinois at Urbana-Champaign (UIUC).
He is Director of the International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL).
He is Principal Investigator on the Networked Environment for Music Analysis project (NEMA).
He has been very active in the establishment of the Music Information Retrieval (MIR) community
through his ongoing work with the International Symposium on Music Information Retrieval (ISMIR) conferences
as a member of the ISMIR steering committee.
He holds a BA (Music Theory and Composition) along with a Master's and a PhD in Library and Information Science,
all earned at the University of Western Ontario, London, Canada.
Donald Byrd studied music composition at Indiana University in the late 1960's,
and then became interested in computers and their potential to help musicians.
After spending a number of years as a programmer and consultant at the University's academic computing support services,
he received a PhD in Computer Science with a dissertation on music notation by computer.
Since then, He has worked extensively both in industry and academia.
He was one of the principal sound designers and sound-design software developers for the Kurzweil 250,
arguably the first synthesizer to reproduce sounds of acoustic instruments convincingly.
He was also the principal designer of the influential music-notation program Nightingale.
His academic background includes research on music notation by computer (at Princeton University);
work on information retrieval in text, especially visualization and human/computer interaction aspects
(at the University of Massachusetts); and work on music information retrieval, digital music libraries,
and optical music recognition (at the University of Massachusetts and Indiana University).
Most recently, he has been working on the "General Temporal Workbench," a timeline-based system for visualizing,
exploring, creating, and "playing" temporal phenomena:
a system general enough for use on any timescale from fractions of an attosecond to billions of years.
He is currently senior scientist and adjunct associate professor in the School of Informatics at IU.
Tim Crawford is a member of the Intelligent Sound and Music Systems group in the Computing Department at Goldsmiths College,
University of London.
He worked for 15 years as a professional musician before turning to academic research.
He is active as a musicologist, being internationally recognized as a leading authority
on the history and music of the European lute,
and is currently Editor of the Complete Works of the lutenist Silvius Leopold Weiss (1687-1750).
Otherwise he is mostly engaged in the application of computational methods to music-related research.
He managed the UK effort for the original OMRAS project (Online Music Recognition and Searching, 1999-2003),
which was the precursor of the currently-running OMRAS2 project on which he currently works.
He also conceived, led and managed ECOLM (Electronic Corpus of Lute Music, 1999-2006),
and is currently Principal Investigator of the Purcell Plus project
which is investigating the application of eScience in musicology
and the longer-term methodological implications of technology for the discipline.
He is one of the founders of ISMIR and frequent contributor as author or organizer
and has recently jointly edited one of a series of books on "Humanities Computing:
Modern Methods for Musicology: Prospects, Proposals and Realities," ISBN 978-0-7546-7302-6 (Farnham: Ashgate 2009),
in which several ISMIR authors are represented.
Keynote Talk: Wind Instrument-Playing Humanoid Robots
09:00-10:00 Thursday (October 29)
Speaker:
- Atsuo Takanishi (Department of Modern Mechanical Engineering, Waseda University, Japan)
Abstract:
Even though the market size is still small at this moment,
applied fields of robots are gradually spreading from the manufacturing industry to the others in recent years.
One can now easily expect that applications of robots will expand into the first and the third industrial fields
as one of the important components to support our society in the 21st century.
There also raises strong anticipations in Japan that robots for the personal use will coexist with humans
and provide supports such as the assistance for the housework, care of the aged and the physically handicapped,
since Japan is one the fastest aging societies in the world.
Consequently, humanoid robots and/or animaloid robots have been treated as subjects of robotics researches in Japan
such as a research tool for human/animal science, an entertainment/mental-commit robot or an assistant/agent for humans
in the human living environment.
Over the last couple of years, some manufactures including famous global companies started to develop prototypes
or even to sell mass production robots for the purposes mentioned above, such as HONDA, TOYOTA, Mitsubishi Heavy, TMSUK, etc.
On the other hand, Waseda University, where the author belongs to,
has been one of the leading research sites on humanoid robot research
since the late Prof. Ichiro Kato and his colleagues started the WABOT (WAseda roBOT) Projects
and developed the historical humanoid robots that were WABOT-1 and WABOT-2 in the early 70s and 80s respectively.
One of the most important aspects of our research philosophy is as follows:
By constructing anthropomorphic/humanoid robots that functions and behaves like a human,
we are attempting to develop the design method of humanoid robots to coexist with humans naturally and symbiotically,
as well as to scientifically build not only the physical model of a human but also its mental model from the engineering view point.
Based upon the philosophy, I and my colleagues have been developing the flute playing humanoid robots as WF (Waseda Flutist) series
as well as the bipedal walking robots WABIAN series, the emotion expression robots WE series and the talking robots WT series, etc.
Especially, the purpose of the flute playing robot research is to build the model of the human flute play and to clarify the model
from the engineering viewpoint by reproducing the human-like flute play using a humanoid robot having the human-like respiratory
organs for the flute play. By using the robot, we become able to experimentally confirm the model of the human flute play quantitatively.
The flute playing robot/model is useful for the flute playing beginners to show how to use/move the organs or it will be used
for the evaluation of the flute instrument production in industry.
We also started the development of saxophone playing humanoid robots recently.
In my keynote talk, I will introduce the research philosophy of my humanoid robots in general by showing examples,
the technical aspects of the wind instrument playing humanoid robots and the other humanoid robots related to music.
Biography
Atsuo Takanishi is a Professor of the Department of Modern Mechanical
Engineering, Waseda University and a concurrent Professor and one of the core
members of the HRI (Humanoid Robotics Institute), Waseda University. He received
the B.S.E. degree in 1980, the M.S.E. degree in 1982 and the Ph.D. degree in
1988, all in Mechanical Engineering from Waseda University.
His current researches are related to Humanoid Robots and its applications in
medicine and well-being, such as the biped walking robots for modeling human
biped walking as WABIAN (WAseda BIpedal humANoid) series, the biped locomotors
for carrying handicapped or elders as WL (Waseda Leg) series, the mastication
robots WJ (Waseda Jaw) series to mechanically simulate human mastication for
clarifying the hypotheses in dentistry, the jaw opening-closing trainer robots
WY (Waseda Yamanashi) series for patients having difficulties in jaw opening or
closing, the flute-playing robots as WF (Waseda Flutist) series and the
saxophone-playing robots WS (Waseda Saxophonist) series to quantitatively
analyze human flute/saxophone playing by collaborating with a professional
flutists/saxophonists, and the anthropomorphic talking robots WT (Waseda Talker)
series which mechanically speak Japanese vowels and consonant sounds, and the
other robots/systems related to his research area. His interest in humanoid
robots has extended to the emotion of human that he started to develop the
emotion expression humanoid robots WE (Waseda Eye) series and KOBIAN/HABIAN
which emotionally behave like a human based upon the "Equations of Emotion." His
humanoid robot WABIAN-2R was exhibited in the 2005 World Exposition in Aichi,
Japan to demonstrate the knee extended walking using the human-like pelvis and
seven DOF leg mechanisms. The emotion expression humanoid KOBIAN is developed
based on WABIAN-2R. The latest model WL-16 carries humans and virtually any
heavy load weighing up to 80 kg. This project is aiming at developing a
practical personal vehicle which supports the society of Japan rapidly becoming
an aging society. He recently developed suture/ligature evaluation system WKS
series which shows surgeon trainees the quantitative scores of their
suture/ligature skills. This system is commercially available from a medical
model and training simulator company, Kyoto Kagaku Co. Ltd., in Japan. He is
also developing the airway management robot for anesthetist/paramedic trainees
collaborating with the company. Refer to www.takanishi.mech.waseda.ac.jp for
more details.
He is a member of Robotics Society of Japan (a board member in 1992 and 1993),
Japanese Society of Biomechanisms, Japanese Society of Mechanical Engineers,
Japanese Society of Instrument and Control Engineers and Society of Mastication
Systems (a major board member from 1996 to current), IEEE and other medicine and
dentistry related societies in Japan.
He received the Best Paper Award from Robotic Society Japan (RSJ) two times in
1998 and in 2005, the Finalist of Best Paper Award two times in the IEEE
International Conference on Robotics and Automation (ICRA) in 1999 and in 2006,
the Best of Asia Award from BusinessWeek Magazine in 2001, the Distinguished
Research Activity Award in Robotics and Mechatronics from Japan Society of
Mechanical Engineers (JSME) in 2003, the Best Paper Award ? Application in
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) in
2004, the Excellent Research Award in 2005 from the Japan Society for Artificial
Intelligence (JSAI), the Industrial Application Division Promotion Award in 2005
from the Society of Instrument and Control Engineers (SICE), the Best Paper
Award in 2006 from JSME, the Best Conference Paper Award in IEEE/ASME
International Conference on Advanced Intelligent Mechatronics (AIM) in 2009, etc.
Panel Discussion: Industrial Panel Discussion
15:45-17:15 Thursday (October 29)
In this panel practitioners from industry will discuss how their
companies are currently using music information retrieval (MIR)
techniques to solve problems for their customers. Panelists will also
discuss emerging areas of MIR research that are particularly relevant
for commercial applications. Audience members will have the opportunity
to ask questions of the panelists.
Panel Organizer:
- Paul Lamere (The Echo Nest, USA)
Panelists:
- Tom Butcher (Microsoft, USA)
- Norman Casagrande (Last.fm, UK)
- Òscar Celma (Barcelona Music and Audio Technologies, Spain)
- Markus Cremer (Gracenote, USA)
- Keiichiro Hoashi (KDDI R&D Laboratories, Japan)
- Kunio Kashino (NTT Laboratories, Japan)
- Malcolm Slaney (Yahoo, USA)
Biographies of the Panelists:
Paul Lamere is the Director of Developer Community at The Echo Nest, a
research-focused music intelligence startup that provides music information
services to developers and partners through a data mining and machine
listening platform. Paul is especially interested in hybrid music
recommenders and using visualizations to aid music discovery.
Tom Butcher joined Microsoft in 2006 to build large-scale web services for
computing and delivering media experiences. His interests include digital
media, artificial intelligence, the Internet, and various intersections
thereof. Currently, Tom is a senior engineer in the Zune group at Microsoft
creating data-driven media experiences, which include automatic playlist
generation, social discovery, recommendations, and more. Prior to joining
Zune, Tom's work encompassed automatic tagging, indexing, and
recommendations at MSN Video. An avid music enthusiast, Tom records
electronic music in his spurious free time using the moniker Codebase.
Norman Casagrande joined last.fm in 2006 as the head of music research.
Since then he has been working on a wide range of problems, including
collaborative filtering for user/item similarity and recommendation, dealing
with scalability, dynamic playlist generation, users insight, audio and
semantic analysis, fingerprint, spam fighting, and many other related
topics.
Òscar Celma is the Chief Innovation Officer at Barcelona Music and Audio
Technologies (BMAT), a spin-off of the Music Technology Group (MTG). BMAT
offers solutions for music discovery and recommendation, musical edutainment
and music copyright detection.
In 2008, Òscar obtained his Ph.D. in Computer Science and Digital
Communication, in the Pompeu Fabra University (Barcelona, Spain). Òscar
worked in the MTG from 2000 till 2008 as a Researcher and Program
Manager. In 2006, he received the 2nd prize in the International Semantic
Web Challenge for the system named "Foafing the Music", a personalized music
recommendation and discovery application.
Markus Cremer joined Fraunhofer Institute for Integrated Circuits (IIS) in
1996 after graduating from Friedrich-Alexander University in Erlangen,
Germany, where he contributed to the design of embedded audio codec
architectures and digital radio broadcast systems. In 2000, Cremer
co-founded the department Metadata at the Fraunhofer Institute for Digital
Media Technology in Ilmenau, Germany. Since 2005, he has been directing
Gracenote's Media Technology Lab in Emeryville, California. Cremer is a
member of IEEE, ACM, and AES, respectively.
Keiichiro Hoashi joined KDDI R&D Laboratories in 1997. His main research
interest is in the area of content-based multimedia information analysis
and retrieval, namely music, images, and video. Currently, he is working
to implement multimedia content analysis technologies in practical
applications and services. He is also working on research projects in
data mining, and recommender systems. He was a lecturer at Waseda
University from 2002 and 2005, and has received his Dr. Eng. degree from
Waseda University in 2007.
Kunio Kashino is Distinguished Technical Member, Supervisor, leading
Media-search Research Team at NTT Communication Science Laboratories,
and Visiting Professor at National Institute of Informatics (NII),
Japan. His team has been working on audio and video analysis, search,
retrieval, and recognition algorithms. Its activities include
development of basic theories as well as their commercial applications
such as Internet content monitoring. He received his PhD from
University of Tokyo for his work on "music scene analysis" in 1995.
Malcolm Slaney is a principal scientist at Yahoo! Research
Laboratory. He received his PhD from Purdue University for his work
on computed imaging. He is a coauthor, with A. C. Kak, of the IEEE
book "Principles of Computerized Tomographic Imaging." This book was
recently republished by SIAM in their "Classics in Applied
Mathematics" Series. He is coeditor, with Steven Greenberg, of the
book "Computational Models of Auditory Function."
Before Yahoo!, Dr. Slaney has worked at Bell Laboratory,
Schlumberger Palo Alto Research, Apple Computer, Interval Research
and IBM's Almaden Research Center. He is also a (consulting)
Professor at Stanford's CCRMA where he organizes and teaches the
Hearing Seminar. His research interests include auditory modeling
and perception, multimedia analysis and synthesis, compressed-domain
processing, music similarity and audio search, and machine learning.
For the last several years he has lead the auditory group at the
Telluride Neuromorphic Workshop.