Llar virtual d'en David Garcia Garzon (Vokimon)
@masterthesis { garciaThesis_AudioAppPrototyping,
	author = {García, D.},
	title = {Rapid Prototyping for Audio Applications},
	year = {2007},
	institution = {Univesitat Pompeu Fabra},
	URL = {files/publications/81359b-BolognaMasterThesis-dgarcia-07.pdf},
	abstract = {The goal of the proposed master thesis is to define an architecture that enables visual prototyping of real-time audio applications and plug-ins. Visual prototyping means that a developer can build a working application, including user interface and processing core, just by assembling elements together and changing their properties in a visual way. Specifically, this research will address the problem of having to bind interactive user interface to a real-time processing core, when both are defined dynamically, the set of components is extensible, it allows bidirectional communication of arbitrary types of data between the interface and the processing core, and, it still fulfils the real-time requirements of audio applications.

@inproceedings { arumi_DataFlowPatternLanguage,
	title = {A Data Flow Pattern Language for Audio and Music Computing},
	author = {Arumí, P. and Garcia, D. and Amatriain, X.},
	year = {2006},
	abstract = {This article describes a set of patterns the authors have seen emerging during years of experience developing assorted applications in the sound and music domain and receiving influences from theoretical models, existing systems, and colleagues. 
Those patterns aim at offering a generative pattern language that falls within a generic data flow architecture. Regardless the audio domain origins of the patterns, we foresee that they have applicability in other domains. Contributions of this paper are General Data-flow Patterns, that address problems about how to organize high-level aspects of the data-flow architecture, by having different types of modules connections; Flow Implementation Patterns, that address how to physically transfer tokens from one module to another, according to the types of flow defined by the "general data-flow patterns''. Tokens life-cycle, ownership and memory management are recurrent issues in those patterns; and finally, Network Usability Patterns, that address how humans can interact with data-flow networks.}, URL = {files/publications/2ddcef-PLOP-2006-arumi-garcia-amatriain.pdf} } @inproceedings { amatriain_ACMMMCLAMFramework, author = {Amatriain, X. and Arumí, P. and Garcia, D.}, title = {CLAM A Framework for Efficient and Rapid Development of Cross-platform Audio Applications}, year = {2006}, abstract = {CLAM is a C++ framework that offers a complete development and research platform for the audio and music domain. Apart from offering an abstract model for audio systems, it also includes a repository of processing algorithms and data types as well as a number of tools such as audio or MIDI input/output. All these features can be exploited to build cross-platform applications or to build rapid prototypes to test signal and media processing algorithms and systems. The framework also includes a number of stand-alone applications that can be used for tasks such as audio analysis/synthesis, plug-in development or metadata annotation.
In this article we give a brief overview of CLAM's features and applications.}, URL = {files/publications/0cbbb2-ACM-MM-2006-xamat.pdf} } @inproceedings { garcia_ExtraccioAcordsAnnotator, title = {Extracció d'acords amb l'Anotador de Música de CLAM}, year = {2006}, abstract = {Aquest article presenta una eina que hem desenvolupat al Grup de Tecnologia Musical de la Universitat Pompeu Fabra. Es tracta d'una eina extensible d'extracció i edició d'informació de la música anomenada CLAM Music Annotator La darrera versió incorpora un extractor d'acords que amplia l'interés de l'eina més enllà de l'ús científic pel que estava pensat. Aquest extractor d'acords pot ser molt útil per als músics i aficionats que fins ara havien de treure els acords de les cançons d'oïda amb prova i error. L'article explica els principis de funcionament del programa i com es pot adaptar a diferents escenaris. També s'explica el cas concret de l'extracció d'acords, es fa una discussió tècnica del disseny, tecnologies emprades i lliçons apresses i, finalment, es comenta quin sera el camí probable que pendrà l'eina en el futur.}, URL = {files/publications/fitxer}, author = {Garcia, D. and Arumí, P. and Amatriain, X.} } @inproceedings { arumi_Testfarm, title = {Testfarm, una eina per millorar el desenvolupament del programari lliure}, year = {2006}, abstract = {En aquest article presentem una nova eina anomenada testfarm que permet automatitzar les tasques de construcció i testeig de programari i monitorizar el seu estat (per exemple apunt per release, falla un test, no compila un programa...). Expliquem les seves funcionalitats contrastant-les amb eines similars i posem de relleu la seva utilitat en diferents escenaris d'ús. A continuació exposem les decisions de disseny més rellevants i la metodologia àgil que s'ha seguit per portar a terme el seu desenvolupament. Per finalitzar, posem testfarm dins el context del desenvolupament de programari lliure, tot intentant respondre com aquesta eina (i altres similars) pot ajudar a millorar-ne el procés.}, URL = {files/publications/fitxer}, author = {Arumí, P. and Sordo, M. and Garcia, D. and Amatriain, X.} } @inproceedings { cano_IndurstrialContentBasedMusicRecommendation, title = {An Industrial-Strength Content-based Music Recommendation System}, journal = {28th Annual International ACM SIGIR Conference}, year = {2005}, address = {Salvador, Brazil}, abstract = {We present a metadata free system for the interaction with massive collections of music, the MusicSurfer. MusicSurfer automatically extracts descriptions related to instrumentation, rhythm and harmony from music audio signals. Together with efficient similarity metrics, the descriptions allow navigation of multimillion track music collections in a flexible and efficient way without the need of metadata or human ratings. }, URL = {files/publications/3ac0d3-SIGIR05-pcano.pdf}, author = {Cano, P. and Koppenberger, M. and Wack, N. and G. Mahedero, J. and Masip, J. and Celma, O. and Garcia, D. and Gómez, E. and Gouyon, F. and Guaus, E. and Herrera, P. and Massaguer, J. and Ong, B. and Ramírez, M. and Streich, S. and Serra, X.} } @inproceedings { cano_contentBasedMusioRecommendation, title = {Content-based Music Audio Recommendation}, journal = {ACM Multimedia}, year = {2005}, address = {Singapore}, abstract = {We present the MusicSurfer, a metadata free system for the interaction with massive collections of music. MusicSurfer automatically extracts descriptions related to instrumentation, rhythm and harmony from music audio signals. Together with efficient similarity metrics, the descriptions allow navigation of multimillion track music collections in a flexible and efficient way without the need for metadata nor human ratings. }, URL = {files/publications/a4924b-ACMMM05-Cano-Koppenberger-Wack.pdf}, author = {Cano, P. and Koppenberger, M. and Wack, N. and G. Mahedero, J. and Aussenac, T. and Marxer, R. and Masip, J. and Celma, O. and Garcia, D. and Gómez, E. and Gouyon, F. and Guaus, E. and Herrera, P. and Massaguer, J. and Ong, B. and Ramírez, M. and Streich, S. and Serra, X.} } @inproceedings { herrera_Mucosa, title = {Mucosa a music content semantic annotator}, year = {2005}, abstract = {MUCOSA (Music Content Semantic Annotator) is an environment for the annotation and generation of music metadata at different levels of abstraction. It is composed of three tiers an annotation client that deals with micro-annotations (i.e. within-file annotations), a collection tagger, which deals with macro-annotations (i.e. across-files annotations), and a collaborative annotation subsystem, which manages large-scale annotation tasks that can be shared among different research centres. The annotation client is an enhanced version of WaveSurfer, a speech annotation tool. The collection tagger includes tools for automatic generation of unary descriptors, invention of new descriptors, and propagation of descriptors across sub-collections or playlists. Finally, the collaborative annotation subsystem, based on Plone, makes possible to share the annotation chores and results between several research institutions. A collection of annotated songs is available, as a `starter pach' to all the individuals or institutions that are eager to join this initiative.}, URL = {files/publications/189652-ISMIR-2005-Herrera.pdf}, author = {Herrera, P. and Celma, O. and Massaguer, J. and Cano, P. and Gómez, E. and Gouyon, F. and Koppenberger, M. and Garcia, D. and G. Mahedero, J. and Wack, N.} } @misc { Amatriain_Annotator, title = {The CLAM Annotator A Cross-platform Audio Descriptors Editing Tool}, year = {2005}, abstract = {This paper presents the CLAM Annotator tool. This application has been developed in the context of the CLAM framework and can be used to manually edit any previously computed audio descriptors. The application offers a convenient GUI that allows to edit low-level frame descriptors, global descriptors of any kind and segmentation marks. It is designed in such a way that the interface adapts itself to a user-defined schema, offering possibilities to a large range of applications.}, URL = {files/publications/9317d2-ismir2005-clam-annotator.pdf}, author = {Amatriain, X. and Massaguer, J. and Garcia, D. and Mosquera, I.} } @inproceedings { celma_mpeg7Tool, title = {Tools for Content-Based Retrieval and Transformation of Audio Using MPEG-7 The SPOffline and the MDTools}, year = {2004}, abstract = {In this workshop we will demonstrate three applications for content-based retrieval and transformations of audio recordings. They illustrate diverse aspects of a common framework for music content description and structuring implemented using the MPEG-7 standard. MPEG-7 descriptions can be generated either manually or automatically and are stored in an XML database. Retrieval services are implemented in the database. A set of musical transformations are defined directly at the level of musically meaningful MPEG-7 descriptors and are automatically mapped onto low-level audio signal transformations. Topics included in the presentation are (1) Description generation procedure, manual annotation of editorial description the MDTools, automatic description of audio recordings, the SPOffline; (2) Retrieval functionalities, local retrieval SPOffline, remote retrieval web-based retrieval; and (3) Transformation utilities the SPOffline.}, URL = {files/publications/AES25-ocelma.pdf}, author = {Celma, O. and Gómez, E. and Janer, J. and Gouyon, F. and Herrera, P. and Garcia, D.} } @inproceedings { arumi_clam, title = {CLAM, Una llibreria lliure per Audio i Música}, year = {2003}, abstract = {CLAM és un framework lliure i orientat a objectes en C++ que ofereix als desenvolupadors solucions de disseny i un conjunt de components reusables per construir aplicacions musicals i d'audio i per la recerca en l'àmbit del processat del senyal. Algunes d'aquestes eines, també lliures, ja s'han desenvolupat per part de l'MTG. La metodologia de desenvolupament de CLAM assegura la seva qualitat. Degut, sobretot, a la incorporació de CLAM a diverses distribucions de GNU/Linux està facilitant l'aparició d'eines multimèdia de tecnologia avançada en entorns lliures.}, URL = {files/publications/jsl03-clam.pdf}, author = {Arumí, P. and Garcia, D. and Amatriain, X.} } @article { garcia02_clamXml, title = {Suport de XML/MPEG-7 per una llibreria de processat d'àudio i música.}, year = {2002}, note = {Software implementations}, publisher = {Enginyeria La Salle}, abstract = {La introducció o de l'estàndard MPEG7, que es refereix a la codificació o en XML de descriptors multimèdia, fa que sigui útil dotar a una llibreria de processat d'àudio i música de la capacitat de treballar amb XML. L'objectiu d'aquest treball és dissenyar, per a una llibreria d'aquest tipus, un sistema activació i passivació, en XML, de les seves dades de processament, i, habilitar l'ús d'XML per altres funcionalitats com ara definició o de sistemes, configuració de processos i represa de càlculs posposats. S'ha volgut minimitzar la feina a fer pels usuaris de la llibreria i els creadors de nous objectes, automatitzant la programació, però, tot donant la possibilitat de personalitzar els formats de sortida. }, URL = {files/publications/salle02-dgarcia.pdf}, author = {Garcia, D.} } @inproceedings { amatriain02_clamFramework, title = {CLAM An OO Framework for Developing Audio and Music Applications}, year = {2002}, abstract = {CLAM (C++ Library for Audio and Music) is a framework for audio and music programming. It may be used for developing any type of audio or music application as well as for doing more complex research related with the field. In this paper we introduce the practicalities of CLAM's first release as well as some of the sample application that have been developed within the framework.}, URL = {files/publications/oopsla02-xamatriain1.pdf}, author = {Amatriain, X. and de Boer, M. and Robledo, E. and Garcia, D.} } @inproceedings { garcia01_xmlControlAudio, title = {XML as a means of control for audio processing, synthesis and analysis}, year = {2001}, abstract = {This paper discusses about benefits derived from providing XML support to the component based framework for audio systems that we are developing. XML is used as data format for persistency, visualization and inter-application interface. Direct XML support is a very useful feature for an audio framework because of the popularity of the XML format as data interchange format, and the introduction of MPEG7 standard, an XML based description format for multimedia content. Formatting task has been distributed along the system objects in a compositional way, making easy to format a single object from its parts. The system minimizes the overhead added to a class and the programmer effort to support XML I/O. A default XML implementation has been provided for most of the future data structures, giving the chance to customize it. The system has been designed to be reused with other formats with a minimal impact on the system.}, URL = {files/publications/mosart2001-garcia.pdf}, author = {Garcia, D. and Amatriain, X.} }