--- tags: PhD --- ```mermaid graph LR raw_text[Raw Text Data] --> preprocessing[Preprocessing] preprocessing --> model_training[Model Training] model_training --> output_generation[Output Generation] output_generation --> final_model[Trained LLM] ``` # PhD Structure ## Top-level outline ```mermaid graph LR A[PhD Dissertation] --> B[Ch 1: Introduction] A --> C[Ch 2: Literature Review] A --> D[Ch 3: Research Methodology] A --> E[Ch 4: Results] A --> F[Ch 5: Case Studies] A --> G[Ch 6: Conclusion] A --> H[Ch 7: Appendices] ``` ### Chapter 1: Introduction ```mermaid graph LR A[Introduction] --> B[Background and Motivation] A --> C[Research Problem] A --> D[Research Approach: Questions and Objectives] A --> E[Scope and Delimitations] A --> F[Proposed Methodology] A --> G[Dissertation Structure] A --> H[Conclusion] B --> BA[The importance of scientific communication] B --> BB[Challenges in processing and understanding large volumes of data] B --> BC[Role of MDS in mitigating these challenges] C --> CA[Defining the problem] C --> CB[Challenges and limitations of existing methods] C --> CC[Significance of addressing this research problem] D --> DA[Developing efficient MDS framework] D --> DB[Investigating the use of advanced NLP techniques] D --> DC[Evaluating framework performance] D --> DD[Potential applications/implications of the research] D --> DE[Research Questions and Objectives] E --> EA[Justifying focus on scientific papers] E --> EB[Limitations of research context/data/methodologies] E --> EC[Assumptions and biases] F --> FA[Introduction to main proposed methodolgies] F --> FB[Rationale for these methodologies] F --> FC[Ethical considerations and risks] G --> GA[Overview of chapters and contents] G --> GB[Navigation guidance] H --> HA[Recap of problem/objectives/significance] H --> HB[Expected contributions and impact/ Contrib to knowledge] ``` ### Chapter 2: Literature Review ```mermaid graph LR A[Literature Review] --> B[General Approaches to MDS] A --> C[Transformer-based Approaches to MDS] A --> D[Hybrid Approaches to MDS] A --> E[LLM approaches to MDS] B --> BA[Extractive vs. Abstractive Summarisation] B --> BB[Single-Document Summarisation vs. MDS] B --> BC[Evaluation Metrics for MDS] B --> BD[Choosing Datasets for MDS] C --> CA[Introduction to Transformers] C --> CB[BART-Bidirectional and Auto-Regressive Transformers] C --> CC[BERT-Bidirectional Encoder Representations from Transformers] C --> CD[GPT-Generative Pre-trained Transformer] C --> CE[XLNet-eXtreme Multi-task Learning via Adversarial Training of a Transformer Network] D --> DA[Introduction to Hybrid Approaches] D --> DB[Combinatorial Approaches] D --> DC[Iterative Approaches] D --> DD[Reinforcement Learning-based Approaches] D --> DE[Transformer-based Hybrid Approaches] E --> EA[Introduction to LLMs] E --> EB[Pre-training Strategies for LLMs] E --> EC[Fine-tuning Strategies for LLMs] E --> ED[Ethical and Societal Implications of LLMs] ``` ### Updated section headings 1. **Introduction to Literature Review** - Brief overview of the chapter - Explanation of the relevance of the reviewed literature to your research topic 2. **Background on Document Summarization** - Overview and history of document summarisation - The importance of document summarisation in various fields - Introduction to single document vs. multi-document summarisation 3. **Extractive vs. Abstractive Summarisation** - Detailed description and comparison of extractive and abstractive summarisation techniques - Advantages and disadvantages of each approach - Explanation of why abstractive summarisation is more relevant to your research 4. **Multi-Document Summarisation (MDS)** - Introduction to MDS and its complexities - Review of key methodologies and techniques used in MDS 5. **Challenges in MDS** - Discussion on the challenges in MDS - Explanation of pre-trained language models like BERT, RoBERTa, GPT-3/4, and T5 6. **Pre-trained Language Models in Summarisation** - Review of studies utilising these models for summarisation tasks - Discussion on the effectiveness of these models in the context of MDS 7. **Domain-Specific Summarisation: Focusing on Scientific Papers** - Overview of the specific challenges in summarising scientific papers - Review of previous work on summarising scientific literature - Discussion on the gaps and potential areas for improvement in current methodologies 8. **Application of Advanced NLP Techniques in MDS** - Overview of advanced NLP techniques, such as transfer learning, attention mechanisms, graph-based methods, and domain adaptation - Review of studies applying these techniques in MDS - Discussion on the potential benefits and drawbacks of these techniques in the context of your research 9. **Choosing and Preparing Data for MDS** - Importance of dataset selection in MDS - Techniques for data preprocessing and formatting - Discussion on the quality and characteristics of suitable datasets 10. **Choice of Evaluation Metrics** - Overview of the common evaluation metrics for summarisation tasks (e.g., ROUGE, BLEU) - Discussion on their strengths and limitations - Explanation of why specific metrics are chosen for this research 11. **Summary of Literature Review** - Recap of the major findings from the literature review - Identification of the gaps in the literature that your research aims to fill ```mermaid digraph attention rankdir=LR; node [shape=record]; input [label="Input Embeddings"]; scaleddotproduct [label="Scaled Dot-Product Attention"]; multihead [label="Multi-Head Attention"]; ffnn [label="Feed-Forward Neural Networks"]; output [label="Output Embeddings"]; input -> scaleddotproduct [label="Q, K, V"]; scaleddotproduct -> multihead [label="Multiple Attention Outputs"]; multihead -> ffnn [label="Concatenated Outputs"]; ffnn -> output [label="Final Output"]; ```