Archive for the ‘Invited Talks’ Category

Talk: August 27, 2015: Effective exploitation of long-term correlations for audio coding and networking

Wednesday, August 26th, 2015

The Department of Electrical Engineering
IEEE Signal Processing Society, Bangalore Chapter

cordially invite you to a lecture on

Title: Effective exploitation of long-term correlations for audio coding
and networking

Speaker: Dr. Tejaswi Nanjundaswamy, Postdoc fellow, UC Santa Barbara

Venue: Multimedia Classroom (MMCR), EE Department

Date and Time: August 27, 2015; 4.30 PM to 5.30 PM.

Abstract: A wide range of multimedia applications such as internet radio
and television, online media streaming, gaming, and high fidelity
teleconferencing heavily rely on efficient transmission of audio signals
over networks. The two main challenges for such transmission is delay
constrained compression, and dealing with loss of content due to noisy
channels. Constraints on delay means that the algorithms can only
operate on small block sizes (or frame lengths). Thus the key to
addressing these challenges is efficiently exploiting inter-frame
redundancies due to long term correlations. While well known audio
coders are effective in eliminating redundancies within a block of data,
and the only known inter-frame redundancy removal technique of employing
a long term prediction (LTP) filter is too simplistic, as it is
suboptimal for the commonly occurring polyphonic audio signals, which
contain a mixture of several periodic components, and also suboptimal
for speech and vocal content, which is quasi-periodic with small
variations in pitch period. Moreover the typically employed parameter
estimation technique is mismatched to the ultimate perceptual distortion
criteria of audio coding. Similarly even in loss concealment, none of
the existing techniques are designed to overcome the main challenge due
to the polyphonic nature of most music signals. This talk covers
contributions towards addressing all these shortcomings by employing
novel sophisticated filter structures suitable for a wide variety of
audio signals, with parameter estimation which takes into account the
perceptual distortion criteria for audio compression, and utilizes all
the available information for loss concealment.

Biography of the speaker: Tejaswi Nanjundaswamy received his B.E degree in electronics and
communications engineering from the National Institute of Technology
Karnataka, India, in 2004, the M.S. and the PhD. degree in electrical
and computer engineering from UCSB, in 2009 and 2013, respectively. He
is currently a post-doctoral researcher at the Signal Compression Lab in
UCSB, where he focuses on audio/video compression, processing and
related technologies. He worked at Ittiam Systems, Bangalore, India from
2004 to 2008 as Senior Engineer in the Audio group. He won the Student
Technical Paper Award at the Audio Engineering Society’s 129th
Convention. He was also a Best Paper Award Finalist at IEEE Workshop on
Applications of Signal Processing to Audio and Acoustics (WASPAA) 2011
and co-author of a Top 10% Award winning paper at IEEE International
Conference on Image Processing (ICIP) 2015.

Talk: May 29, 2015: Bayesian Models for Computational Rhythm Analysis in Indian Art Music

Saturday, May 16th, 2015

Department of Electrical Communication Engineering and
Department of Electrical Engineering,
Indian Institute of Science, Bangalore

invite you to a talk by
Ajay Srinivasamurthy
Music Technology Group, Universitat Pompeu Fabra
Barcelona, Spain

Bayesian Models for Computational Rhythm Analysis in Indian Art Music

Time & Date: 11 AM, May 29, 2015.
Venue: Golden Jubilee Hall, Department of Electrical Communication Engineering (ECE)

Rhythm in Indian art music (Carnatic and Hindustani music) is organized in the framework of tāla (or tāl). A tala consists of time cycles that provide a broad structure for repetition of music phrases, motifs and improvisations. Detecting different events such as the beats and the sama (downbeats) within a tāla cycle and tracking them through a piece of audio recording, referred to as meter inference, is an important computational rhythm analysis task in Indian Art Music. Useful in itself, it additionally provides a basis for further computational analyses such as structural analysis and extraction of rhythmic patterns. The talk mainly focuses on my recent work with Bayesian models for meter inference in Carnatic music. Efficient approximate inference in these models are presented to overcome the limitations of exact inference. Further,I will discuss extensions of these models that generalize to other music styles and different metrical structures.

Biography of the Speaker
Ajay is a PhD student at the Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain. He is a part of the CompMusic project led by Prof. Xavier Serra, where he works on rhythm related research problems in Indian Art Music and Beijing Opera. He is currently working towards developing signal processing and machine learning approaches for the analysis and characterization of rhythmic structures and patterns from audio recordings. Prior to joining UPF, he was a research assistant at the Georgia Tech Center for Music Technology, Atlanta, USA and worked at the mobile music startup Smule Inc. He has a masters in Signal Processing from the Indian Institute of Science, Bangalore, India and a B.Tech in Electronics and Communication Engineering from National Institute of Technology Karnataka, Surathkal, India.

Co-sponsors of the event:
IEEE Bangalore section
IEEE SP Society Bangalore Chapter

Special Lecture: 10 March 2015: Our recent work related to the design of filter banks with uncertainty considerations

Monday, March 9th, 2015







Dear All:

Speaker: Prof. V. M. Gadre, IIT Bombay, Mumbai, India
Title: Our recent work related to the design of filter banks with uncertainty considerations

Date/time: Tuesday, 10th March 2015, 3pm (coffee/tea at 2.45pm)
Venue: ECE GJH

Abstract: Two band filter banks have typically been designed with passband and stopband considerations and there is scant work on designing them with the uncertainty criterion. In this talk, we present some of the work that we have done in designing the two band filter bank with the uncertainty principle as a criterion.

Bio: Prof. Vikram Gadre received his Ph. D in Electrical Engineering from Indian Institute of Technology (IIT) Delhi in 1994, and thereafter has been working at IIT Bombay. He is currently the head of Centre for Distance Engineering Education. His research areas are communication and signal processing. He has taught 16 courses like signals and systems, control systems. He received ‘Excellence in Teaching’ award from IIT Bombay in 1999, 2004, and 2009. He has guided sponsored research projects for organizations like Tata Infotech, Texas Instruments, and 9 PhD, 65 M.Tech and Dual Degree, and 45 B.Tech projects. He has authored two books and has published 60 conference proceedings, 23 journal papers, and chapters in edited monographs.


Talks: 18 December 2014

Sunday, February 15th, 2015

IEEE Signal Processing Society, Bangalore Chapter, IEEE Bangalore Section,
and Supercomputer Education Research Centre, Indian Institute of Science

Invite you to the following talks:

1) Title: Beyond Mindless Labeling: *Really* Leveraging Humans to Build
Intelligent Machines

by Dr. Devi Parikh, Assistant Prof., Virginia Tech.


Human ability to understand natural images far exceeds machines today. One
reason for this gap is an artificially restrictive learning set up –
humans today teach machines via Morse code (e.g. providing binary labels
on images, such as “this is a horse” or “this is not”), and machines are
typically silent. These systems have the potential to be significantly
more accurate if they tap into the vast common-sense knowledge humans have
about the visual world. I will talk about our work on enriching the
communication between humans and machines by exploiting mid-level visual
properties or attributes. I will also talk about the more difficult
problem of directly learning common-sense knowledge simply by observing
the structure of the visual world around us. Unfortunately, this requires
automatic and accurate detection of objects, their attributes, poses, etc.
in images, leading to a chicken-and-egg problem. I will argue that the
solution is to give up on photorealism. Specifically, I will talk about
our work on exploiting human-generated abstract visual scenes to learn
common-sense knowledge and study high-level vision problems.


2) Title: Hedging Against Uncertainty in Machine Perception via Multiple
Diverse Predictions

by Dr. Duruv Bhatra, Assistant Prof., Virginia Tech.


What does a young child or a high-school student with no knowledge of
probability do when faced with a problem whose answer they are uncertain
of? They make guesses.

Modern machine perception algorithms (for object detection, pose
estimation, or semantic scene understanding), despite dealing with
tremendous amounts of ambiguity, do not.

In this talk, I will describe a line of work in my lab where we have been
developing machine perception models that output not just a single-best
solution, rather a /diverse/ set of plausible guesses. I will discuss
inference in graphical models, connections to submodular maximization over
a “doubly-exponential” space, and how/why this achieves state-of-art
performance on challenging Pascal VOC 2012 segmentation dataset.


SPECIAL LECTURE: 17 October 2014 : Design of Sparse Antenna Arrays

Wednesday, February 11th, 2015






Dear All,

Title: Design of Sparse Antenna Arrays

Speaker: Prof. Sanjit Mitra, Univ. of California, Santa Barbara, and Univ.
of Southern California

Date: 17 October 2014 (Friday)

Time: 11:00am (tea/coffee at 10:45am)

Venue: Golden Jubilee Hall, ECE Dept., IISc Bangalore

Abstract: In this talk we present a general method for the factorization
of a polynomial with unity coefficients into a product of sparse
polynomials with fewer non-zero coefficients with unity values. The
factorization method is then used to design sparse antenna arrays with
uniform and linearly tapered effective aperture functions.

Bio: Sanjit K. Mitra is a Research Professor in the Department of
Electrical & Computer Engineering, University of California, Santa Barbara
and Professor Emeritus, Ming Hsieh Department of Electrical Engineering,
University of Southern California, Los Angeles.

He has held visiting appointments in Australia, Austria, Brazil, Croatia,
Finland, Germany, India, Japan, Norway, Singapore, Turkey, and the United

Dr. Mitra has published over 700 papers in the areas of analog and digital
signal processing, and image processing. He has also authored and
co-authored twelve books, and holds five patents. He has presented 29
keynote and/or plenary lectures at conferences held in the United States
and 16 countries abroad. Dr. Mitra has served IEEE in various capacities
including service as the President of the IEEE Circuits & Systems Society
in 1986.

Dr. Mitra is the recipient of the 1973 F.E. Terman Award and the 1985 AT&T
Foundation Award of the American Society of Engineering Education; the
1989 Education Award, and the 2000 Mac Van Valkenburg Society Award of the
IEEE Circuits & Systems Society; the Distinguished Senior U.S. Scientist
Award from the Alexander von Humboldt Foundation of Germany in 1989; the
1996 Technical Achievement Award, the 2001 Society Award and the 2006
Education Award of the IEEE Signal Processing Society; the IEEE Millennium
Medal in 2000; the McGraw-Hill/Jacob Millman Award of the IEEE Education
Society in 2001; the 2002 Technical Achievement Award and the 2009
Athanasios Papoulis Award of the European Association for Signal
Processing, the 2005 SPIE Technology Achievement Award of the
International Society for Optical Engineers; the University Medal of the
Slovak Technical University, Bratislava, Slovakia in 2005; the 2006 IEEE
James H. Mulligan, Jr. Education Medal; and the 2013 IEEE Gustav Robert
Kirchhoff Award. He is the corecipient of the 2000 Blumlein-Browne-Willans
Premium of the Institution of Electrical Engineers (London) and the 2001
IEEE Transactions on Circuits & Systems for Video Technology Best Paper
Award. He has been awarded Honorary Doctorate degrees from the Tampere
University of Technology, Finland, the Technical University of Bucharest,
Romania, and the Technical University of Iasi, Romania.

He is a member of the U.S. National Academy of Engineering, a member of
the Norwegian Academy of Technological Sciences, an Academician of the
Academy of Finland, a foreign member of the Finnish Academy of Sciences
and Arts, a corresponding member of the Croatian Academy of Sciences and
Arts, and the Academy of Engineering, Mexico, and a Foreign Fellow of the
National Academy of Sciences, India and the Indian National Academy of
Engineering. Dr. Mitra is a Fellow of the IEEE, AAAS, and SPIE.

All are invited

SPECIAL LECTURE: May 29th, 2014 : Frequency modulated continuous wave radars: A gentle introduction to angle resolution and its challenges

Wednesday, February 11th, 2015







Title:  Frequency modulated continuous wave radars: A gentle introduction to angle resolution and its challenges
Date: Thursday, May 29th, 2014
Time: 4pm
Tea/coffee: 3.45pm
Venue: Golden Jubilee Hall, ECE Department, IISc Bangalore
Speaker: Sandeep Rao, Texas Instruments India
Abstract :
77GHz FMCW radars have been used in the automotive and niche industrial applications for a while now. However as technology improves and the levels of integration increase, there is a tremendous potential for these radars to be used in wide variety of other applications. As the use of radar becomes more wide-spread, its design becomes more cost-sensitive. One of the key design aspects which contributes to the cost of an FMCW radar is the number of antennas. Hence the motivation to get the maximum performance out of a small number of antennas. The first part of this talk will be a survey of some relevant aspects of angle estimation. We will highlight the importance  of multi-resolution techniques and discuss some related algorithms (MUSIC, ML based). We will also discuss techniques for improving angle resolution through appropriate array design.  In this context we will survey the concepts of  Active beam-forming, passive beam-forming and   synthetic aperture radar. The second part of the talk will focus on angle estimation specifically in the context of low cost FMCW radar systems. The interdependence between angle resolution, range resolution and velocity resolution will be discussed. We will discuss the open challenges in achieving acceptable angle resolution with a minimum number of non-ideal antennas. These challenges span multiple areas : antenna design, angle resolution algorithms, blind and non-blind calibration techniques, detection algorithms and sensor fusion. Application specific challenges will also be discussed.

Sandeep Rao is with the systems group in the wireless division at Texas Instruments and currently looks at system and algorithmic aspect of automotive radars. Earlier he led the GPS positioning algorithm development effort at TI. Prior to TI, he was with Hughes Network Systems where he was involved in the development of modem algorithms for satellite communications. He has a Masters from the University of Maryland and a Bachelors from the Indian Institute of Technology Madras.

All are invited

SPECIAL LECTURE: May 26th, 2014 : Frequency modulated continuous wave radars: an introduction and mm-wave chip design challenges

Wednesday, February 11th, 2015







Title:  Frequency modulated continuous wave radars: an introduction and mm-wave chip design challenges
Date: Monday, May 26th, 2014
Time: 4pm
Tea/coffee: 3.45pm
Venue: Golden Jubilee Hall, ECE Department, IISc Bangalore
Speaker: Karthik Subburaj, Texas Instruments India
Abstract :
Frequency Modulated Continuous Wave RADAR systems are popular in automotive and industrial applications to measure relative distance, velocity and direction of surrounding objects. In the first part of this talk, the signal processing fundamentals of a few RADAR methods, starting from Doppler RADAR and leading up to the modern FMCW RADARs are introduced. In the latter part of the talk, various challenges in implementing them using Integrated Circuits and in improving performance wrt range, accuracy and false object detection are discussed. The challenges include RF and analog non-idealities such as noise and non-linearity, and simultaneous operation of transmit and receive antennas and solving them is crucial in realizing a high performance radar system. The contents of this talk are a mix of the fields of RF system design and basic signal processing.

Karthik Subburaj is with the systems group in wireless division at Texas Instruments, currently focusing on signal processing algorithms and RF system budgeting for radar systems. He has earlier worked on PHY layer signal processing and digital design for GNSS receivers, FM transceivers, Timing Recovery in High Speed Serial Interfaces and digital PLLs. He holds an M.S degree from University of Florida, Gainesville and B.E from College of Engineering, Guindy.

All are invited

Talk: January 9, 2015: Segregating Complex Sound Sources Through Temporal Coherence

Wednesday, February 11th, 2015

Dept. of Electrical Engineering, IISc, invites you to the following talk
related to using auditory processing insights for segregating sound
Title: Segregating Complex Sound Sources Through Temporal Coherence
Speaker: Prof. Shihab A. Shamma,
Univ. of Maryland, USA and Ecole Normale Superieure, Paris
Time & Date: 4:00 pm; Friday January 9, 2015
Venue: Multimedia Classroom (PE 217), Department of Electrical
Engineering, IISc.

Abstract: A new approach for the segregation of monaural sound mixtures
is presented based on the principle of temporal coherence and using
auditory cortical representations. Temporal coherence is the notion that
perceived sources emit coherently modulated features that evoke
highly-coincident neural response patterns. By clustering the feature
channels with coincident responses and reconstructing their input, one can
segregate the underlying source from the simultaneously interfering
signals that are uncorrelated with it. The proposed algorithm requires no
prior information or training on the sources. It can, however, gracefully
incorporate cognitive functions and influences such as memories of a
target source or attention to a specific set of its attributes so as to
segregate it from its background. Aside from its unusual structure and
computational innovations, the proposed algorithmic model provides
testable hypotheses of the physiological mechanisms of this ubiquitous and
remarkable perceptual ability, and of its psychophysical manifestations in
navigating complex sensory environments.

Biography of the Speaker:
Shihab Shamma received his B.S. degree in 1976 from Imperial College, in
London, U.K. He received his M.S. and Ph.D. degrees in Electrical
Engineering from Stanford University in 1977 and 1980, respectively. Dr.
Shamma received his M.A. in Slavic Languages and Literature in 1980 from
the same institution. Dr. Shamma has been a member of the University of
Maryland faculty since 1984 when he started as an Assistant Professor in
the Electrical Engineering Department. He became an Associate Professor in
1989. He has been associated with the Systems Research Center since its
inception in 1985, and received a joint appointment in 1990. Dr. Shamma
also holds a joint appointment with the University of Maryland Institute
for Advanced Computer Studies. He is a fellow of the Acoustical Society of
America since 2004. He received the ISR Outstanding Systems Engineering
Faculty Award from University of Maryland in the year 2007. His research
interests include representation of the acoustic signal at various levels
in mammalian auditory systems and ranges from theoretical models of
auditory processing in early and central auditory stages, to
neuro-physiological investigations of the auditory cortex, to
psycho-acoustical experiments of human perception of acoustic spectral

Dr. Shamma’s research deals with issues in computational neuroscience,
neuromorphic engineering, and the development of micro sensor systems for
experimental research and neural prostheses. Primary focus has been on
studying the computational principles underlying the processing and
recognition of complex sounds (speech and music) in the auditory system,
and the relationship between auditory and visual processing. Signal
processing algorithms inspired by data from neuro-physiological and psycho
acoustical experiments are being developed and applied in a variety of
systems such as speech and voice recognition and diagnostics in industrial
manufacturing. Other research interests included (at various times) the
development of photo-lithographic micro-electrode arrays for recording and
stimulation of neural signals, a VLSI implementations of auditory
processing algorithms, and development of robotic systems for the
detection and tracking of multiple sound sources.
IEEE Signal Processing Society
Bangalore chapter

Talk: September 8, 2014 : Time-Frequency Modulation Analysis of Autoregressive Spectrogram Models

Wednesday, February 11th, 2015

IEEE Signal Processing Society, Bangalore Chapter, IEEE Bangalore Section and
Department of Electrical Engineering, IISc

invite you to a talk by
Dr. Sriram Ganapathy
IBM T.J. Watson Research Center,
Yorktown Heights, USA

Time-Frequency Modulation Analysis of Autoregressive Spectrogram Models

Time & Date: 4:00pm; Monday September 8, 2014
Venue: Multimedia Classroom (PE 217), Department of Electrical Engineering, IISc.

Even with several advancements in the practical application of speech technology, the performance of speech systems remain fragile in high levels of noise and other environmental distortions. On the other hand, various studies on the human auditory system have shown good resilience of the system to high levels of noise and degradations. This information shielding property of the auditory system may be largely attributed to the signal peak preserving functions performed by the cochlea and the spectro-temporal modulation filtering
performed by the cortical stages. In this talk, I will discuss how these steps can be efficiently emulated in the front-end signal processing of automatic speech systems. A spectrographic representation that emphasizes the high energy regions is derived using time-frequency
autoregressive modeling followed by a modulation filtering step to preserve only the key spectro-temporal modulations. This talk will conclude with experimental results from several interesting speech applications.

Biography of the Speaker
Sriram Ganapathy is a research staff member at the IBM T.J. Watson Research Center, Yorktown Heights, USA, where he is working on signal analysis methods for radio communication speech in highly degraded environments. He received his Doctor of Philosophy from the Center for Language and Speech Processing, Johns Hopkins University in 2011 with Prof. Hynek Hermansky. He obtained his Bachelor of Technology from College of Engineering, Trivandrum, India in 2004 and Master of Engineering from the Indian Institute of Science, Bangalore in 2006. In the past, he has also been part of the Idiap Research Institute in Martigny, Switzerland from 2006 to 2008 contributing to various speech and Audio projects. His research interests include signal processing, machine learning and robust methodologies for speech and speaker recognition. He has secured over 50 publications in leading international journals and conferences in speech and audio processing along with several patents.

Talk: December 15, 2014 : Beyond Siri: Towards the Next Generation of Talking and Listening Machines

Wednesday, February 11th, 2015

Department of Electrical Engineering, Indian Institute of Science
IEEE Signal Processing Society, Bangalore Chapter

invite you to a talk by
Prof. Roger K. Moore
University of Sheffield, U.K.

Beyond Siri: Towards the Next Generation of Talking and Listening Machines
at 3:30pm; December 15, 2014; Venue: Multimedia Classroom (PE 217), Department of Electrical Engineering, IISc.

Over the past thirty years, the field of spoken language processing has made impressive progress from simple laboratory demonstrations to mainstream consumer products. However, the limited abilities of commercial applications such as Siri highlight the fact that there is still some way to go in creating Autonomous Social Agents that are truly capable of conversing effectively with their human counterparts in real-world situations. What seems to be missing is an overarching theory of intelligent interactive behaviour that is capable of informing the system-level design of such systems. This talk addresses these issues and argues that we need to go far beyond our current capabilities and understanding towards a more integrated perspective; we need to move from developing devices that simply talk and listen to evolving intelligent communicative machines that are capable of truly understanding human behaviour.

Biography of the Speaker
Prof. Roger K. Moore studied Computer and Communications Engineering at the University of Essex and was awarded the B.A. (Hons.) degree in 1973. He subsequently received the M.Sc. and Ph.D. degrees from the same university in 1975 and 1977 respectively, both theses being on the topic of automatic speech recognition. After a period of post-doctoral research in the Phonetics Department at University College London, Prof. Moore was head-hunted in 1980 to establish a speech recognition research team at the Royal Signals and Radar Establishment (RSRE) in Malvern.

In 1985 Prof. Moore became head of the newly created ‘Speech Research Unit’ (SRU) and subsequently rose to the position of Senior Fellow (Deputy Chief Scientific Officer – Individual Merit) in the ‘Defence and Evaluation Research Agency’ (DERA). Following the privatisation of the SRU in 1999, Prof. Moore continued to provide the technical lead as Chief Scientific Officer at 20/20 Speech Ltd. (now Aurix Ltd.) – a joint venture company between DERA (now QinetiQ) and NXT plc. In 2004 Prof. Moore was appointed Professor of Spoken Language Processing in the ‘Speech and Hearing’ Research Group (SPandH) at Sheffield University, where he is pioneering research that is aimed at developing computational models of spoken language processing by both mind and machine. Prof. Moore is currently working on a unified theory of spoken language processing in the general area of ‘Cognitive Informatics’ called ‘PRESENCE’ (PREdictive SENsorimotor Control and Emulation). PRESENCE weaves together accounts from a wide variety of different disciplines concerned with the behaviour of living systems – many of them outside the normal realms of spoken language – and compiles them into a new framework that is intended to breathe life into a new generation of research into spoken language processing.

Prof. Moore has authored and co-authored over 150 scientific publications in the general area of speech technology applications, algorithms and assessment. He is a Fellow of the UK Institute of Acoustics, a Visiting Professor in the Department of Phonetics and Linguistics at University College London and a Visiting Professor at the Bristol Robotics Laboratory. He is Editor-in-Chief of ‘Computer Speech and Language’ and a member of the Editorial/Advisory boards for ‘Speech Communication’, ‘Languages’ and the ‘International Journal of Cognitive Informatics and Natural Intelligence’ (IJCiNi). He is past Chairman of the ‘European Expert Advisory Group on Language Engineering Standards’ (EAGLES) working party on spoken language resources, and Editor of the ‘Handbook of Standards and Resources for Spoken Language Systems’. Prof. Moore served as President of the ‘European Speech Communication Association’ (ESCA) and the ‘International Speech Communication Association’ (ISCA) from 1997 to 2001, and as President of the Permanent Council of the ‘International Conferences on Spoken Language Processing’ (PC-ICSLP) from 1996 to 2001. During this period he pioneered the internationalisation of ESCA, the integration of the EUROSPEECH and ICSLP conferences into an annual INTERSPEECH conference, and chaired the joint ISCA/PC-ICSLP working party which drew up the detailed recommendations for the merger. In 1994 Prof. Moore was awarded the prestigious UK Institute of Acoustics Tyndall medal for “distinguished work in the field of speech research and technology” and in 1999 he was presented with the NATO RTO Scientific Achievement Award for “repeated contribution in scientific and technological cooperation”. In 2008 he was elected as one of the first ISCA Fellows “in recognition of his applications of human speech perception and production models to speech technologies and his service to ISCA as President”. Prof. Moore was General Chair for INTERSPEECH 2009.