In this study, two outstanding subgroups of organic-inorganic hybrid materials have been investigated. The first part covers the design, synthesis, characterization and application of seven novel Metal Organic Frameworks (MOFs) containing functionalized biphenyl dicarboxylates as linkers. In the second part, the surface modification of the metal oxides ZrO2, TiO2 and Al2O3 using phosphonate derivates is reported.
Firstly three functionalized MOF structures; ZnBrBPDC, ZnNO2BPDC and ZnNH2BPDC were synthesised using 4,4´-biphenyldicarboxylic acid derivatives with different functional groups (-Br, -NO2, -NH2) Powder X-ray diffraction (PXRD) measurements indicated that the synthesised MOFs posses the interpenetrated IRMOF-9 structure with a cubic topology, which was also confirmed with single crystal X-ray measurements. The chemical structure of the MOF materials was further proved by solid state NMR and IR measurements. N2 adsorption measurements showed Type I isotherms for all three structures with large surface areas. TGA measurements of the evacuated samples were in good agreement with the elemental analysis data. The results proved that their thermal stability is between 325 °C - 450 °C.
Adsorption properties of these MOF structures were tested using light alkanes (CH4, C2H6, C3H8, and n-C4H10) at three different temperatures. For all adsorbents, the maximum uptakes were observed at 273 K. When the temperature was increased, the amount of the adsorbed gas decreased. All three MOFs showed strong affinities for n-butane. The lowest uptakes were observed for CH4.
The effect of functional groups on the IRMOF series was also examined by synthesizing amide functionalized biphenyl linkers. For this purpose, four different linkers containing amides with different alkyl chains (C1-C4) were synthesized and used for the synthesis of four new MOF structures ZnAcBPDC, ZnPrBPDC, ZnBuBPDC and ZnPeBPDC.
PXRD measurements of ZnAcBPDC indicated that the structure contains two different phases. PXRD patterns of ZnPrBPDC, ZnBuBPDC and ZnPeBPDC revealed non-interpenetrated structures which were further proved by single crystal X-ray measurements. The chemical structure of the MOF materials was further confirmed by X-ray spectoscopy, solid state NMR and IR measurements.
N2 adsorption measurements of the MOF structures were carried out using different activation methods. For all four MOFs, Type I isotherms were obtained. ZnAcBPDC showed the highest BET surface area. ZnAcBPDC and ZnBuBPDC were tested for their alkane, alkene and CO2 adsorption capacities.
In the second part of the work, the surface modification of three different metal oxides, ZrO2, TiO2 and Al2O3 was performed. For this purpose firstly three different fluorescent phosphonate derivatives containing thiophene units were synthesized from their halo derivatives in a four step synthesis and then used as coupling molecules for the surface modification. Nine different surfaces were obtained (38@TiO2, 39@TiO2, 40@TiO2, 38@Al2O3, 39@Al2O3, 40@Al2O3, 38@ZrO2, 39@ZrO2, 40@ZrO2).
All three modified metal oxide surfaces were characterized using elemental analysis, solid state NMR and IR spectroscopy. The BET surface areas of the materials were determined by N2 adsorption measurements. TGA was used to determine the stability of the surfaces. Maximum loadings were obtained for ZrO2 surfaces.
Due to the strong luminescence of the coupling molecules, the modified surfaces were checked for their light emission. All ZrO2 and Al2O3 surfaces showed fluorescence with exception of 40@Al2O3. On the other hand, for the modified TiO2 surfaces, no fluorescence could be observed.

I report on two experiments, which were designed to test theoretical predictions about individual behavior in a duopolistic setting. With quantity being the choice variable a simultaneous Cournot game and a sequential Stackelberg game were tested over two periods. The key feature of both models was that players were able to lower marginal cost for period two if they successfully outperformed their competition in period one in terms of profit. Experimental results suggest that in the Cournot game players are very competitive in period one but become Cournot players in period two. In the Stackelberg game Cournot play is modal, suggesting that players have preferences for equality in payoffs, which maybe brought about by punishment of Stackelberg followers and fear of punishment of Stackelberg leaders . Overall, players earned more money in the Stackelberg game than in the Cournot game.

Generic layout analysis--process of decomposing document image into homogeneous regions for a collection of diverse document images--has many important applications in document image analysis and understanding such as preprocessing of degraded warped, camera-captured document images, high performance layout analysis of document images containing complex cursive scripts, and word spotting in historical document images at page level. Many areas in this field like generic text line extraction method are considered as elusive goals so far, still beyond the reach of the state-of-the-art methods [NJ07, LSZT07, KB06]. This thesis addresses this problem in such a way that it presents generic, domain-independent, text line extraction and text and non-text segmentation methods, and then describes some important applications, that were developed based on these methods. An overview of the key contributions of this thesis is as follows.
The first part of this thesis presents a generic text line extraction method using a combination of matched filtering and ridge detection techniques, which are commonly used in computer vision. Unlike the state-of-the-art text line extraction methods in the literature, the generic text line extraction method can be equally and robustly applied to a large variety of document image classes including scanned and camera-captured documents, binary and grayscale documents, typed-text and handwritten documents, historical and contemporary documents, and documents containing different scripts. Different standard datasets are selected for performance evaluation that belong to different categories of document images such as the UW-III [GHHP97] dataset of scanned documents, the ICDAR 2007 [GAS07] and the UMD [LZDJ08] datasets of handwritten documents, the DFKI-I [SB07] dataset of camera-captured documents, Arabic/Urdu script documents dataset, and German calligraphic (Fraktur) script historical documents dataset. The generic text line extraction method achieves 86% (n = 23,763 text lines in 650 documents) text line detection accuracy which is better than the aggregate accuracy of 73% of the best performing domain-specific state-of-the-art methods. To the best of the author's knowledge, it is the first general-purpose text line extraction method that can be equally used for a diverse collection of documents.
This thesis also presents an active contour (snake) based curled text line extraction method for warped, camera-captured document images. The presented approach is applied to DFKI-I [SB07] dataset of camera-captured, Latin script document images for curled text line extraction. It achieves above 95% (n = 3,091 text lines in 102 documents) text line detection accuracy, which is significantly better than the competing state-of-the-art curled text line extraction methods. The presented text line extraction method can also be applied to document images containing different scripts like Chinese, Devanagari, and Arabic after small modifications.
The second part of this thesis presents an improved version of the state-of-the-art multiresolution morphology (Leptonica) based text and non-text segmentation method [Blo91], which is a domain-independent page segmentation approach and can be equally applied to a diverse collection of binarized document images. It is demonstrated that the presented improvements result in an increase in segmentation accuracy from 93% to 99% (n = 113 documents).
This thesis also introduces a discriminative learning based approach for page segmentation, where a self-tunable multi-layer perceptron (MLP) classifier [BS10] is trained for distinguishing between text and non-text connected components. Unlike other classification based page segmentation approaches in the literature, the connected components based discriminative learning based approach is faster than pixel based classification methods and does not require a block segmentation method beforehand. A segmentation accuracy of $96\%$ ($n = 113$ documents) is achieved in comparison to the state-of-the-art multiresolution morphology (Leptonica) based page segmentation method [Blo91] that achieves a segmentation accuracy of 93%. In addition to text and non-text segmentation of Latin script documents, the presented approach can also be adapted for document images containing other scripts as well as for other specialized layout analysis tasks such as digit and non-digit segmentation [HBSB12], orientation detection [RBSB09], and body-text and side-note segmentation [BAESB12].
Finally, this thesis presents important applications of the two generic layout analysis techniques, ridge-based text line extraction method and the multi-resolution morphology based text and non-text segmentation method, discussed above. First, a complete preprocessing pipeline is described for removing different types of degradations from grayscale warped, camera-captured document images that includes removal of grayscale degradations such as non-uniform shadows and blurring through binarization, noise cleanup applying page frame detection, and document rectification using monocular dewarping. Each of these preprocessing steps shows significant improvement in comparison to the analyzed state-of-the-art methods in the literature. Second, a high performance layout analysis method is described for complex Arabic script document images written in different languages such as Arabic, Urdu, and Persian and different styles for example Naskh and Nastaliq. The presented layout analysis system is robust against different types of document image degradations and shows better performance for text and non-text segmentation, text line extraction, and reading order determination on a variety of Arabic and Urdu document images as compared to the state-of-the-art methods. It can be used for large scale Arabic and Urdu documents' digitization processes. These applications demonstrate that the layout analysis methods, ridge-based text line extraction and the multi-resolution morphology based text and non-text segmentation, are generic and can be applied easily to a large collection of diverse document images.

The main topic of this thesis is to define and analyze a multilevel Monte Carlo algorithm for path-dependent functionals of the solution of a stochastic differential equation (SDE) which is driven by a square integrable, \(d_X\)-dimensional Lévy process \(X\). We work with standard Lipschitz assumptions and denote by \(Y=(Y_t)_{t\in[0,1]}\) the \(d_Y\)-dimensional strong solution of the SDE.
We investigate the computation of expectations \(S(f) = \mathrm{E}[f(Y)]\) using randomized algorithms \(\widehat S\). Thereby, we are interested in the relation of the error and the computational cost of \(\widehat S\), where \(f:D[0,1] \to \mathbb{R}\) ranges in the class \(F\) of measurable functionals on the space of càdlàg functions on \([0,1]\), that are Lipschitz continuous with respect to the supremum norm.
We consider as error \(e(\widehat S)\) the worst case of the root mean square error over the class of functionals \(F\). The computational cost of an algorithm \(\widehat S\), denoted \(\mathrm{cost}(\widehat S)\), should represent the runtime of the algorithm on a computer. We work in the real number model of computation and further suppose that evaluations of \(f\) are possible for piecewise constant functions in time units according to its number of breakpoints.
We state strong error estimates for an approximate Euler scheme on a random time discretization. With this strong error estimates, the multilevel algorithm leads to upper bounds for the convergence order of the error with respect to the computational cost. The main results can be summarized in terms of the Blumenthal-Getoor index of the driving Lévy process, denoted by \(\beta\in[0,2]\). For \(\beta <1\) and no Brownian component present, we almost reach convergence order \(1/2\), which means, that there exists a sequence of multilevel algorithms \((\widehat S_n)_{n\in \mathbb{N}}\) with \(\mathrm{cost}(\widehat S_n) \leq n\) such that \( e(\widehat S_n) \precsim n^{-1/2}\). Here, by \( \precsim\), we denote a weak asymptotic upper bound, i.e. the inequality holds up to an unspecified positive constant. If \(X\) has a Brownian component, the order has an additional logarithmic term, in which case, we reach \( e(\widehat S_n) \precsim n^{-1/2} \, (\log(n))^{3/2}\).
For the special subclass of $Y$ being the Lévy process itself, we also provide a lower bound, which, up to a logarithmic term, recovers the order \(1/2\), i.e., neglecting logarithmic terms, the multilevel algorithm is order optimal for \( \beta <1\).
An empirical error analysis via numerical experiments matches the theoretical results and completes the analysis.