Fachbereich IV
Refine
Year of publication
Document Type
- Doctoral Thesis (30)
- Book (4)
- Habilitation (3)
- Report (3)
Has Fulltext
- yes (40)
Is part of the Bibliography
- no (40)
Keywords
- Machine Learning (3)
- Argumentation (1)
- Bewertungskompetenz (1)
- Blaulichtemission (1)
- Case-based Reasoning (1)
- Case-based decision support (1)
- Clustering (1)
- Complexity Metrics (1)
- Data Mining (1)
- Decentralized Model Training (1)
Institute
The EXPLAIN project (EXPLanatory interactive Artificial intelligence for INdustry) aims at enabling explainable Machine Learning in industry. MLOps (Machine Learning Operations) includes tools, practices, and processes for deploying ML (Machine Learning) in production. These will be extended by explainability methods as part of the project.
This study aims to determine to what extent MLOps is implemented by four project partner companies. Further, the study describes the ML use cases, MLOps software architecture, tools, and requirements in the companies perspective. Besides, requirements for a novel MLOps software architecture, including explainability methods, are collected.
As a result the interviews show that each of the interviewed industry partners use MLOps differently. Different tools and architectural patterns are used depending on the particular use case. Overall, most information we gathered focused on architecture decisions in the MLOps tool landscape used by the interviewed companies.
The discourse about pathways of sustainability transitions is dominated by two opposing approaches. On the one hand, the transition studies literature follows a green growth efficiency-driven approach and argues that technological innovations will pave the way to more sustainable modes of living. On the other hand, postcapitalist scholars reject the efficiency paradigm and advocate an alternative, sufficiency-driven turn to sustainability characterized by degrowth-oriented downscaling and rightsizing.
This habilitation thesis contributes to this field of research and investigates how transitions towards more sustainable paths can be initiated and fostered. It focuses on tourism in rural regions that have been growing intensely in recent years and are characterized by mass tourism causing ecological and socio-cultural pressure. The thesis thus analyses what influence particularly tourists and tourism businesses have on the production of rural space regarding sustainability and rural tourism resilience and what role rural regions may generally play in sustainability transition processes.
The study recognizes that tourists and tourism businesses have a strong influence on the production of rural space while their sustainability orientations are heterogeneous. Rural tourism resilience and sustainable change are challenged by the growing popularity of rural tourism and the still dominantly growth-oriented rural tourism sector. Future rural development paths will be positioned within a spectrum of green growth-oriented mass tourism on the one hand and degrowth-oriented tourism on the other hand. With respect to shared responsibilities of urban and rural regions, this thesis pleads for financial compensation as a degrowth-oriented incentive for those rural tourism regions that are willing to follow a concept of reduced use of economic resources for the sake of greater ecological and socio-cultural goals that are of value for our societies on different geographical scales.
The University of Hildesheim owns many different software systems for teaching. So far,
no one has attempted to cover all systems and create an architectural description of e-learning at the University of Hildesheim. This report describes the applications relevant to education. These include software developed by the university, third-party tools provided by the university, cloud tools closely associated with the university, and tools provided by the university under contract. Implemented functionalities are explored,
and functional specifcations are described, including information about technical and operational requirements. Here, the software is primarily teaching tools. However, the report does not include software and development environments in which students are taught but rather tools that apply to teaching. The variety of tools in different courses makes it impossible to cover all the applications that are used by work groups or individuals.
The main goal of this investigation is to capture the university's e-learning state and describe the architecture - mainly to ease the development of applications for developers.
Further objectives are to identify new ideas and summarize current goals.
We conducted surveys of the teaching staff to determine the current status. Transcripts of interviews, tables provided by interviewed persons, and websites were used as references to prepare this document. We hope that future investigations will provide updates on current information and add information in this context.
In dieser Arbeit werden die Ergebnisse eines Design-Based Research Projekts zur Didaktik der Gruppentheorie berichtet. Grundlage aller diesbezüglichen Erhebungen bildet das aus der Literatur entwickelte Hildesheimer Lehrkonzept zur Gruppentheorie, welches sowohl die didaktische Perspektive der Neuen Mathematik als auch moderne Forschungserkenntnisse berücksichtigt. Ziel des Lehrkonzepts besteht dabei in der Heranführung Lernender zu einem adäquaten Konzeptverständnis gruppentheoretischer Grundlagen.
Das Hildesheimer Lehrkonzept wurde dazu im Sinne des Design-Based-Research Pardigmas in unterschiedlichen
Forschungszyklen evaluiert und überarbeitet – zunächst auf Grundlage einer formativen Evaluation mittels Akzeptanzbefragung und anschließend im Rahmen einer summativen Evaluation mit
N = 143 Lehramtsstudierenden im ersten Semester. Für die summative Evaluation im Prätest-Posttest-Design wurde mit dem CI2GT ein neu entwickeltes Testinstrument zur Erhebung des Konstrukts Konzeptverständnis Gruppentheorie eingesetzt, das im Rahmen dieser Arbeit auch einer ausführlichen psychometrischen
Charakterisierung unterzogen wurde, sowohl mithilfe klassischer als auch probabilistischer Testtheorie.
Das Mixed Methods Design, also die Synthese qualitativer und quantitativer Studien, sowie die Untersuchung kognitiver und affektiver Lernermerkmale, erlauben einen vielseitigen Blick auf didaktische Aspekte der Gruppentheorie, der gleichzeitig auch das Herzstück dieser Arbeit bildet. Die Ergebnisse zeichnen folgendes Bild: Lernende erleben Gruppentheorie als eine äußerst relevante Teildisziplin der Mathematik. Lernschwierigkeiten zeigen sich dabei vor allem in sprachlich bedingten Präkonzepten und dem axiomatischen Fundament. Die Arbeit schließt mit einem Ausblick in verwandte Themenbereiche, in denen sprachlich bedingte Präkonzepte ebenfalls beobachtet werden.
Road accidents are one of the leading causes of death worldwide, particularly among young people. The police and local authorities therefore strive to reduce the risk of accidents through appropriate road safety measures. In order to plan these measures, the relevant types of accidents, i. e., accidents with certain features, must first be recognized. However, the variety of accident features and the amount of resulting feature combinations make it impossible to monitor all accident types manually.
In this thesis, methods are proposed to automatically identify interesting accident types. Here, it is investigated whether combinations of features occur together and how the temporal pattern of the combined occurrence behaves. The change mining approach can then be used to determine whether structural changes in frequency occur during the period under consideration. For example, a feature combination that suddenly appears more frequently or exhibits a change in seasonality should be prioritized for further investigation so that appropriate road safety measures may be initiated for that combination.
The implemented strategic, multi-stage data mining framework based on frequent itemset mining, time series clustering, forecasting methods, and a scoring process is able to detect interesting feature combinations. These are then processed on a map in a web interface suitable for the respective audience in order to support the strategic planning of road safety measures. The framework is applied to several accident data sets from different countries to determine suitable default parameter values for the respective data analysis methods and to carefully align the methods. It is shown that there exist only minor dependencies of the parameter selection on the database to be analyzed.
For operational planning, it is necessary to consider small geographic areas and identify the features that have the greatest impact on accident occurrence there. Therefore, the developed operational framework analyzes and predicts the course of accident time series, taking into account the associated feature-specific time series. On the one hand, this makes it possible to increase the forecast performance, and, on the other hand, to determine which accident features have a significant influence on the course of the accident numbers over time. The insights gained can be used as a basis for short-term measures.
Supervised learning, the standard paradigm in machine learning, only works well if a sufficiently large, diverse, and cleanly-annotated dataset is available. Unfortunately, this is often not the case. In fact, the lack of labeled data is an omnipresent issue in machine learning. The problem is particularly prevalent in computer vision, where unlabeled images or videos can often be acquired at a low cost, whereas labeling them is time-consuming and expensive. To address the issue, this thesis focuses on developing new methods that aim at reducing annotation costs in computer vision by leveraging unlabeled and partially labeled data.
In the first part, we provide an overview of previous research directions and discuss their strengths and weaknesses. Thereby, we identify particularly promising research areas. The subsequent chapters which form the central part of this thesis aim at developing algorithmic improvements in these especially attractive fields. Among them is self-supervised learning, which aims at learning transferable representations given a large number of unlabeled images. We find that existing self supervised methods are optimized for image classification tasks, only compute global per-image feature vectors, and are designed for object-centric datasets like ImageNet. To address these issues, we propose a method that is particularly suited for object detection downstream tasks and works well if multiple objects are present per image like in video data for autonomous driving. Another core downside of self-supervised learning algorithms is that they depend on very large batch sizes with batch norm statistics synchronized across GPUs and also require many epochs of training until convergence. We find that stabilizing the self-supervised training target substantially speeds up convergence and allows for training with much smaller batch sizes. Our method matches ImageNet weights after 25 epochs of training with a batch size of only 32.
Finally, we investigate supervised pretraining. We find that state-of-the-art self-supervised methods match ImageNet weights only in classification or detection but not in both. In addition, we show that more sophisticated supervised training strategies significantly improve upon ImageNet weights.
The second part of the thesis deals with partially labeled data for object detection. We propose to label only large, easy-to-spot objects given a limited budget. We argue that these contain more pixels and therefore usually more information about the underlying object class than small ones. At the same time, they are easier to spot and hence cheaper to label. Because conventional supervised learning algorithms do not work well given this annotation protocol, we develop our own method with does, by combining pseudo-labels, output consistency across scales, and an anchor scale-dependent ignore strategy. Furthermore, many object detection datasets such as MS COCO and CityPersons include group annotations, i.e., bounding boxes that contain multiple objects of a single class. We find that pseudo-labeling instances within a group box is superior to the commonly used training strategies.
In the third part of the thesis, we cover semi-supervised object detection where a subset of the images is fully labeled whereas the remaining ones are unlabeled. We show that existing methods that are almost exclusively developed for Faster R-CNN work much less well if applied to architectures that are sensitive to missing annotations. In the prefinal chapter, we investigate the interaction between data and computer vision algorithms. This is in contrast to the vast majority of research which considers the data to be fixed. We provide computer vision practitioners and researchers with guidelines about what to do in typical situations.
In the final part of the thesis, we discuss the overall findings and investigate if research should put greater weight on acquiring and labeling data. Finally, we discuss options of mimicking human learning with machines, which might eventually result in human-level intelligence. After all, humans are living proof that this kind of learning works, if done properly.
Recent decades have seen exponential growth in data acquisition attributed to advancements in edge device technology. Factory controllers, smart home appliances, mobile devices, medical equipment, and automotive sensors are a few examples of edge devices capable of collecting data. Traditionally, these devices are limited to data collection and transfer functionalities, whereas decision-making capabilities were missing. However, with the advancement in microcontroller and processor technologies, edge devices can perform complex tasks. As a result, it provides avenues for pushing training machine learning models to the edge devices, also known as learning-at-the-edge. Furthermore, these devices operate in a distributed environment that is constrained by high latency, slow connectivity, privacy, and sometimes time-critical applications. The traditional distributed machine learning methods are designed to operate in a centralized manner, assuming data is stored on cloud storage. The operating environment of edge devices is impractical for transferring data to cloud storage, rendering centralized approaches impractical for training machine learning models on edge devices.
Decentralized Machine Learning techniques are designed to enable learning-at-the-edge without requiring data to leave the edge device. The main principle in decentralized learning is to build consensus on a global model among distributed devices while keeping the communication requirements as low as possible. The consensus-building process requires averaging local models to reach a global model agreed upon by all workers. The exact averaging schemes are efficient in quickly reaching global consensus but are communication inefficient. Decentralized approaches employ in-exact averaging schemes that generally reduce communication by communicating in the immediate neighborhood. However, in-exact averaging introduces variance in each worker's local values, requiring extra iterations to reach a global solution.
This thesis addresses the problem of learning-at-the-edge devices, which is generally referred to as decentralized machine learning or Edge Machine Learning. More specifically, we will focus on the Decentralized Parallel Stochastic Gradient Descent (DPSGD) learning algorithm, which can be formulated as a consensus-building process among distributed workers or fast linear iteration for decentralized model averaging. The consensus-building process in decentralized learning depends on the efficacy of in-exact averaging schemes, which have two main factors, i.e., convergence time and communication. Therefore, a good solution should keep communication as low as possible without sacrificing convergence time. An in-exact averaging solution consists of a connectivity structure (topology) between workers and weightage for each link. We formulate an optimization problem with the objective of finding an in-exact averaging solution that can achieve fast consensus (convergence time) among distributed workers keeping the communication cost low. Since direct optimization of the objective function is infeasible, a local search algorithm guided by the objective function is proposed. Extensive empirical evaluations on image classification tasks show that the in-exact averaging solutions constructed through the proposed method outperform state-of-the-art solutions.
Next, we investigate the problem of learning in a decentralized network of edge devices, where a subset of devices are close to each other in that subset but further apart from other devices not in the subset. Closeness specifically refers to geographical proximity or fast communication links.
We proposed a hierarchical two-layer sparse communication topology that localizes dense communication among a subgroup of workers and builds consensus through a sparse inter-subgroup communication scheme. We also provide empirical evidence of the proposed solution scaling better on Machine Learning tasks than competing methods.
Finally, we address scalability issues of a pairwise ranking algorithm that forms an important class of problem in online recommender systems. The existing solutions based on a parallel stochastic gradient descent algorithm define a static model parameter partitioning scheme, creating an imbalance of work distribution among distributed workers. We propose a dynamic block partitioning and exchange strategy for the model parameters resulting in work balance among distributed workers. Empirical evidence on publicly available benchmark datasets indicates that the proposed method scales better than the static block-based methods and outperforms competing state-of-the-art methods.
In der hier vorgelegten Promotionsarbeit wird das Potenzial gruppenbasierter und semistrukturierter Aushandlungsprozesse analysiert. In einer entsprechenden Interventionsstudie mit Pre- und Post-Analysen wurden 146 Schüler:innen einer Gesamtschule in Niedersachsen/Deutschland aufgefordert, Begründungen zu acht selbstentwickelten Argumenten zu einem Thema über den Erhalt der lokalen Biodiversität, einem bioethischen Konflikt im Rahmen von nachhaltiger Entwicklung, vor und nach einer gruppenbasierten Aushandlung zu formulieren und diese zu gewichten. Zu diesem Zweck verwendeten die Schüler:innen in allen Phasen die Zielmat als ein Instrument zur Strukturierung des Bewertungsprozesses. Die Begründungen wurden inhaltsanalytisch hinsichtlich der Nutzung argumentativer Ressourcen analysiert. Darüber hinaus wurde die Richtung der Veränderungen der Begründungen nach der Aushandlung qualitativ verglichen und die Veränderung der Gewichtungen quantitativ berechnet. Bei diesen Analysen wurden individuelle Gewichtungen und Begründungen beider Phasen und die Veränderung der Gewichtungen mit den Gruppengewichtungen verglichen. Die Ergebnisse der Begründungen zeigen, dass die Schüler:innen bereits vor dem Aushandlungsprozess über eine Bandbreite an argumentativen Ressourcen (nämlichen faktenbasierte und normative Ressourcen) verfügen. Die Ergebnisse des Vergleichs der Begründungen von der Pre- zur Post-Phase zeigen, dass etwa ein Drittel aller Begründungen verändert wurden. Die Richtung der Veränderung ist zudem sehr divers, da die Schüler:innen die Begründungen widerlegten, revidierten, aber auch bestätigten und verstärkten. Ebenso wurde etwa ein Drittel aller Gewichtungen in der Post-Phase verändert. Ein Vergleich der Gewichtungsänderung der Pre- zu Post-Phase mit der Gruppengewichtung zeigt, dass diese der Tendenz der Gruppengewichtung entspricht.
Die Ergebnisse dieser Studie machen auf das Potenzial gruppenbasierter Aushandlungsprozesse in bioethischen Konflikten aufmerksam, nämlich die Aktivierung relevanter argumentativer Ressourcen und die Initiierung tiefer und revidierender Denkprozesse. Darüber hinaus zeigen die Daten das besondere Potenzial der in der Studie verwendeten Zielmat, nämlich die Unterstützung komplexer und sonst für Schüler:innen überfordernder kompensatorischer Gewichtungsstrategien.
Finding an available parking spot in city centers can be a cumbersome task for individual drivers and also negatively affects general traffic flow and CO2 emissions.
In the context of smart cities and the internet of things this problem can be mitigated by using available data to monitor and predict parking occupancy in order to guide users to an available parking location near their destination.
With this goal in mind there arise multiple challenges of which we introduce selected ones to propose novel solutions based on machine learning.
The focus of this work is to enable the usage of readily available and inexpensive data sources like parking meter transactions, opposed to expensive technology like in-ground sensors or cameras where the costs prevent a widespread coverage. Our proposed data sources do not directly monitor the actual parking availability but still provide enough signal for our algorithms to infer the real parking situation with high accuracy.
As part of this work we developed a parking availability prediction system based on parking meter transactions that was deployed to 33 german cities.
A main contribution of our work is the proposal of a novel way to generate labels based on the parking transactions and to use semi-supervised-, more specifically positive-unlabeled learning, to leverage the sparse signal in order to require as little data as possible.
Additionally, we utilize and design novel methodologies in the area of transfer learning to learn simultaneously from different cities which leads to the previously seldom explored setting of combining transfer learning with positive-unlabeled learning. We therefore introduce a novel algorithm to tackle this problem type.
We hope that our work enables the deployment of smart parking systems at lower costs and therefore leads towards the goal of smart parking guidance in smart cities.
Human skeletal remains are one of several find categories from archaeological sites. The skeleton constitutes only a small part of a former living organism that was exposed to a variety of environmental factors. As a highly adaptive mineralized tissue, bone – and to a lesser extent also teeth – stores information on an individual’s life. Reading out this information contributes to the historical understanding of individuals and societies from the past. However, obtaining this information can be challenging in many aspects, and, due to the very nature of the archaeological remains, the reconstructed picture of the individual and its society and environment will inevitably always be incomplete. In order to extract as much information from bones and teeth as possible, existing methods must be adapted to specific situations and the diagnostic approaches have to be developed beyond current limitations.
The 17 international publications presented in this thesis are addressing the abovementioned goals. They cover a wide range of topics, mainly from the field of bioarchaeology, augmented by some studies on recent and fossil animal remains. The studies are organized in those dealing with metrical data, including sex determination methods and analysis of measurement error, the analysis of normal and abnormal conditions of bone, including paleopathological diagnostics, the impact of diagenetic agents on bone, and the presentation of cases where bioarchaeological analysis contributed to the understanding of the respective find situations. The methods applied in the studies cover a broad spectrum of approaches to study morphological properties of bones and teeth, including metrics, radiographic, and microscopic techniques.