Filters








25 Hits in 1.8 sec

Vergleich von IR-Systemkonfigurationen auf Komponentenebene

Kürsten Jens, Wilhelm Thomas, Eibl Maximilian
2020 Zenodo  
Vergleich von IR-Systemkonfigurationen auf Komponentenebene Jens Kürsten, Thomas Wilhelm und Maximilian Eibl Technische Universität Chemnitz Straße der Nationen 62, 09111 Chemnitz {jens.kuersten, eibl  ...  Experimentaufbau In den vergangenen fünf Jahren wurde ein hochgradig flexibles Retrieval Framework entwickelt, dessen Konzept in (Kürsten 2008 ) dargestellt ist.  ... 
doi:10.5281/zenodo.4134501 fatcat:grkxptyygvar3du64c66qrjwhy

The Xtrieval Framework at CLEF 2008: Domain-Specific Track [chapter]

Jens Kürsten, Thomas Wilhelm, Maximilian Eibl
2009 Lecture Notes in Computer Science  
This article describes our participation at the Domain-Specific track. We used the Xtrieval framework [2], [3] for the preparation and execution of the experiments. The translation of the topics for the cross-lingual experiments was realized with a plug-in to access the Google AJAX language API 2 . This year, we submitted 20 experiments in total. In all our experiments we applied a standard top-k pseudo-relevance feedback algorithm. Also, all of our submissions were merged experiments, where
more » ... tiple stemming approaches for each language were combined to improve retrieval performance. The evaluation of the experiments showed that the combination of stemming methods works very well. Translating the topics for the bilingual experiments deteriorated the retrieval effectiveness only between 8 and 15 percent in comparison to our best monolingual experiments.
doi:10.1007/978-3-642-04447-2_27 fatcat:garlgde5ljhqhna4ousk6negei

Experiments for the ImageCLEF 2007 Photographic Retrieval Task

Thomas Wilhelm, Jens Kürsten, Maximilian Eibl
2007 Conference and Labs of the Evaluation Forum  
This article describes the configuration of the experiments that we submitted for the ImageCLEF Photographic Retrieval Task. We used a redesigned version of our last years retrieval system prototype (see [1] for details). The translation of the topics for our cross-lingual experiments was realized with a plug-in to access the Google Translate [2] service. We used thesauri from OpenOffice [3] to expand the queries for better retrieval performance. This year, we submitted 11 runs, whereof only
more » ... was completely automatic. In all our experiments mixed modality was applied, i.e. we used text retrieval and content-based image retrieval for re-ranking. The evaluation results show that most of our experiments achieved very strong retrieval performance.
dblp:conf/clef/WilhelmKE07 fatcat:ioyak5dmarcvrourtet7nhzi4i

QA Extension for Xtrieval: Contribution to the QAst track

Jens Kürsten, Holger Kundisch, Maximilian Eibl
2008 Conference and Labs of the Evaluation Forum  
This article describes our first participation at the QAst task of the CLEF campaign 2008. We submitted 4 experiments in total, two for each subtask t1 and t4. These subtasks employed manual speech transcription collections. We used the Stanford Named Entity Recognizer for tagging named entities and the CRFTagger -Conditional Random Fields Part-of-Speech (POS) Tagger for English. The passage retrieval was done with the Xtrieval framework and its Apache Lucene implementation. For the
more » ... on of the question hand-crafted patterns were implemented. Our experiments achieved an accuracy of about 20%. The rate of returned NIL answers was too high for all of our experiments.
dblp:conf/clef/KurstenKE08 fatcat:mluo7uwtwvabxioip4q6asyotm

The Xtrieval Framework at CLEF 2008: ImageCLEF Photographic Retrieval Task

Thomas Wilhelm, Jens Kürsten, Maximilian Eibl
2008 Conference and Labs of the Evaluation Forum  
This paper describes our participation at the ImageCLEF photographic retrieval task. We used our Xtrieval framework for the preparation and execution of the experiments. This year, we submitted 4 experiments in total. The experiments showed that our thesaurus based query expansions works well in improving the geometric mean average precision (GMAP) and binary preference (BPREF), but deteriorates the improvements gained by the addition of content-based image retrieval. The baseline (text-only)
more » ... ored a mean average precision (MAP) of 0.0998. The combination of text and image retrieval gained a raise by 37 percent to a MAP of 0.1364. After applying the query expansion to both experiments the MAP for the text-only retrieval increased to 0.1081, but the MAP for the combined text and image retrieval decreased to 0.1140. By implementing an interface to the PostgreSQL database the retrieval speed and comparison operations for vectors could be speeded up.
dblp:conf/clef/WilhelmKE08 fatcat:vyvlinsal5e3vn53znvp52mfjq

The Xtrieval Framework at CLEF 2008: ImageCLEF Wikipedia MM task

Thomas Wilhelm, Jens Kürsten, Maximilian Eibl
2008 Conference and Labs of the Evaluation Forum  
This paper describes our participation at the ImageCLEF Wikipedia MM task. We used our Xtrieval framework for the preparation and execution of the experiments. We submitted 4 experiments in total. The results of these experiments were mixed. The text-only experiment scored second best with a mean average precision (MAP) of 0.2166. In combination with image based features the MAP dropped to 0.2138. With the addition of our thesaurus based query expansion it scored best with a MAP of 0.2195.
more » ... ut query expansion and with the inclusion of the provided concepts the lowest MAP of 0.2048 was achieved, but there were 23 more relevant documents retrieved than in all 3 other experiments. Furthermore, the retrieval speed and comparison operations for vectors could be speeded up by implementing an interface to the PostgreSQL database.
dblp:conf/clef/WilhelmKE08a fatcat:lzbz6ztlyvaingirticqiexlsi

The Importance of being Grid: Chemnitz University of Technology at Grid@CLEF

Maximilian Eibl, Jens Kürsten
2009 Conference and Labs of the Evaluation Forum  
This paper describes the participation of the Chemnitz University of Technology at Grid@CLEF 2009. We integrated the CIRCO framework into our Xtrieval framework and performed 15 runs in the three languages German, English, and French. For each language we used two different stemmers and two different retrieval models. One run one was a fusion run combining the results of the four other experiments. Whereas the different runs demonstrated that the impact of the used retrieval technologies is
more » ... ly depending on the corpus, the merged approach produced the best results in each language.
dblp:conf/clef/EiblK09 fatcat:scbrtgcoarfyrcuoboy3htknmi

Domain-Specific Cross Language Retrieval: Comparing and Merging Structured and Unstructured Indices

Jens Kürsten, Maximilian Eibl
2007 Conference and Labs of the Evaluation Forum  
This year, we participated in all Monolingual, Bilingual and Multilingual tasks of the Domain-Specific track. We used a redesigned version of our retrieval system prototype from 2006, which is based on the Lucene API [1] . A plugin to access the online translation services Google Translate [2] and PROMT [3] was implemented for the cross-language experiments. Furthermore, we tried to figure out the differences between plain and structured indices and also applied a data fusion approach for both
more » ... ndex schemes. In comparison to the median of all participants of the Monolingual tasks we achieved average performance for our german and english and strong performance for our russian runs. The results of the cross-language tasks were robust compared to our own monolingual experiments and better than the average of the results submitted by all participants.
dblp:conf/clef/KurstenE07 fatcat:g3tvcyoolfe5vcq6q2dhhaidya

CLEF 2008 Ad-Hoc Track: On-line Processing Experiments with Xtrieval

Jens Kürsten, Thomas Wilhelm, Maximilian Eibl
2008 Conference and Labs of the Evaluation Forum  
This article describes our first participation at the Ad-Hoc track. We used the Xtrieval framework [2], [3] for the preparation and execution of the experiments. We regard our experiments as online or live experiments since the preparation of all results including indexing and retrieval took us less than 4 hours in total. This year, we submitted 18 experiments in total, whereof only 4 were pure monolingual runs. In all our experiments we applied a standard top-k pseudo-relevance feedback
more » ... hm. The translation of the topics for the multilingual experiments was realized with a plug-in to access the Google AJAX language API 2 . The performance of our monolingual experiments was slightly below the average for the German and French collection and in the top 5 for the English collection. Our bilingual experiments performed very well (at least in the top 3) for all target collections.
dblp:conf/clef/KurstenWE08c fatcat:ll5acarqebadfdmuguw52ept3e

Monolingual Retrieval Experiments with a Domain-Specific Document Corpus at the Chemnitz University of Technology [chapter]

Jens Kürsten, Maximilian Eibl
2007 Lecture Notes in Computer Science  
Abstr act This article describes the first participation of the Media Informatics Section of the Chemnitz Technical University at the Cross Language Evaluation Forum. A first experimental prototype is described which implements several different methods of optimizing search results. The configuration of the prototype is tested with the GIRT corpus. The results of the DomainSpecific Monolingual German task suggest that combining the suffix stripping stemming and the decompounding approach is
more » ... useful. Also, a local document clustering approach used to improve pseudo relevance feedback seems to be quite beneficial. Nevertheless, the evaluation of the English task using the same configuration suggests that the qualities of the results are highly speech dependent.
doi:10.1007/978-3-540-74999-8_26 fatcat:ocwog2mcu5fexi6gcdxq7wqhum

Chemnitz at VideoCLEF 2009: Experiments and Observations on Treating Classification as an IR Task

Jens Kürsten, Maximilian Eibl
2009 Conference and Labs of the Evaluation Forum  
Kürsten et. al. [2] and He et. al. [1] tried to solve the problem with state of the art classifiers like k-NN and SVM. Both used Wikipedia articles to train their classifiers.  ... 
dblp:conf/clef/KurstenE09b fatcat:eyjyxlsqcbb5natce4pwebwnzy

Does Patent IR Profit from Linguistics or Maximum Query Length?

Daniela Becks, Maximilian Eibl, Julia Jürgens, Jens Kürsten, Thomas Wilhelm, Christa Womser-Hacker
2011 Conference and Labs of the Evaluation Forum  
participated together in the CLEF Intellectual Property Track. We focused on the prior art candidate search, which was already provided for the third time. Our group submitted seven runs ranging from simple bag of words to linguistic phrases. The aim of our experiments was to examine the effectiveness of different query strategies. Especially, we wanted to evaluate the advantage of linguistic phrases in contrast to very long bag of words queries. Phrases were extracted using a special
more » ... component, which has been developed by the University of Hildesheim.
dblp:conf/clef/BecksEJKWW11 fatcat:be5ic36lwfb43bqmpnma4dxnju

Chemnitz at CLEF IP 2012: Advancing Xtrieval or a Baseline Hard to Crack

Thomas Wilhelm, Jens Kürsten, Maximilian Eibl
2012 Conference and Labs of the Evaluation Forum  
For the 2012 CLEF-IP Claims to passage task we reused and improved our Xtrieval framework. Our two-step approach comprises creating two Lucene indexes: one containing the whole patent application documents and one containing the same documents split into passages. We prepared three setups and conducted each with a translated and an untranslated topic set, which was just applied to the claims. The submitted setups differ in the way of retrieving the results and merging them. No further
more » ... were used. Therefore our experiments had very simple setups, which nevertheless achieved good results. There are still plenty of possible improvements, which can easily be tested with our framework, because it offers a comprehensive set of methods for conducting and evaluating retrieval experiments.
dblp:conf/clef/WilhelmKE12 fatcat:hqbuegumdrfsfnaa2jrgokpdum

Chemnitz at CLEF 2009 Ad-Hoc TEL Task: Combining Different Retrieval Models and Addressing the Multilinguality

Jens Kürsten
2009 Conference and Labs of the Evaluation Forum  
In this paper we report our efforts for the participation in the CLEF 2009 Ad-Hoc TEL task. In our second participation we were able to test and evaluate a new feature of the Xtrieval framework, which was the accessibility of the three core retrieval engines Lucene, Lemur and Terrier. This year we submitted 24 experiments in total, 12 each for the monolingual and bilingual subtasks. We compared our baseline experiments to combined runs, where we used two different retrieval models, namely the
more » ... ctor space model (VSM) used in Lucene and the Bose-Einstein model for randomness (BB2) available in the Terrier framework. We found that an almost constant improvement in terms of mean average precision over all provided collections is achievable. Furthermore we tried to benefit from the multilingual contents of the collections by running combined multilingual experiments for both subtasks. The evaluation showed that the used approach achieves small improvements in the monolingual setting of the task. Unfortunately, we were not able to confirm this finding in the bilingual setting, where the multilingual experiments were outperformed by the standard bilingual runs, especially on the English target collection.
dblp:conf/clef/Kursten09 fatcat:gu6cnntlj5dc3b3oy7427rsjde

Chemnitz at the CHiC Evaluation Lab 2012: Creating an Xtrieval Module for Semantic Enrichment

Jens Kürsten, Thomas Wilhelm, Daniel Richter, Maximilian Eibl
2012 Conference and Labs of the Evaluation Forum  
Cultural heritage is one of the most valuable resources that describe the creative power of mankind. In this article we describe a total number of 96 experiments that have been submitted as contributions to the three subtasks of the Cultural Heritage in CLEF pilot evaluation lab. At the core of the majority of these experiments lies a prototype implementation for semantic enrichment based on DBpedia. The evaluation of the experiments demonstrate that semantic enrichment does not improve
more » ... l effectiveness in comparison to straightforward baseline experiments. The results also indicate that automatic query expansion does not improve retrieval performance for the pilot lab test collection. Further experiments are needed in order to be able to draw conclusions on whether semantic enrichment can improve retrieval results on cultural heritage collections or not.
dblp:conf/clef/KurstenWRE12 fatcat:gasbgafi3jf25jay53h5a77hgq
« Previous Showing results 1 — 15 out of 25 results