aviles-rivero
    • Create new note
    • Create a note from template
      • Sharing URL Link copied
      • /edit
      • View mode
        • Edit mode
        • View mode
        • Book mode
        • Slide mode
        Edit mode View mode Book mode Slide mode
      • Customize slides
      • Note Permission
      • Read
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Write
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Engagement control Commenting, Suggest edit, Emoji Reply
    • Invite by email
      Invitee

      This note has no invitees

    • Publish Note

      Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

      Your note will be visible on your profile and discoverable by anyone.
      Your note is now live.
      This note is visible on your profile and discoverable online.
      Everyone on the web can find and read all notes of this public team.
      See published notes
      Unpublish note
      Please check the box to agree to the Community Guidelines.
      View profile
    • Commenting
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
      • Everyone
    • Suggest edit
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
    • Emoji Reply
    • Enable
    • Versions and GitHub Sync
    • Note settings
    • Note Insights New
    • Engagement control
    • Make a copy
    • Transfer ownership
    • Delete this note
    • Save as template
    • Insert from template
    • Import from
      • Dropbox
      • Google Drive
      • Gist
      • Clipboard
    • Export to
      • Dropbox
      • Google Drive
      • Gist
    • Download
      • Markdown
      • HTML
      • Raw HTML
Menu Note settings Note Insights Versions and GitHub Sync Sharing URL Create Help
Create Create new note Create a note from template
Menu
Options
Engagement control Make a copy Transfer ownership Delete this note
Import from
Dropbox Google Drive Gist Clipboard
Export to
Dropbox Google Drive Gist
Download
Markdown HTML Raw HTML
Back
Sharing URL Link copied
/edit
View mode
  • Edit mode
  • View mode
  • Book mode
  • Slide mode
Edit mode View mode Book mode Slide mode
Customize slides
Note Permission
Read
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Write
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Engagement control Commenting, Suggest edit, Emoji Reply
  • Invite by email
    Invitee

    This note has no invitees

  • Publish Note

    Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

    Your note will be visible on your profile and discoverable by anyone.
    Your note is now live.
    This note is visible on your profile and discoverable online.
    Everyone on the web can find and read all notes of this public team.
    See published notes
    Unpublish note
    Please check the box to agree to the Community Guidelines.
    View profile
    Engagement control
    Commenting
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    • Everyone
    Suggest edit
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    Emoji Reply
    Enable
    Import from Dropbox Google Drive Gist Clipboard
       Owned this note    Owned this note      
    Published Linked with GitHub
    • Any changes
      Be notified of any changes
    • Mention me
      Be notified of mention me
    • Unsubscribe
    ## Reviewer [hF2P] Apologies for the slow reply. I did not have access to internet in the past week. I think I'm happy with the questions for now. I will however, request the authors to include a more precise description of the functions I asked about, e.g. make it clear that $\Phi$ is a map representing a single image only, with the inputs and outputs specified. Ideally, I would also like to see short description of how the training is done, so it would be replicable even for a non-expert. ## Action Editor [oUSG] The results in the paper seem very promising! I have several questions regarding the paper and hope the authors can help clarify them. Terminology: I am a bit confused by the new terms: order compactness, frequency compactness, and spatial compactness. In Fourier analysis, order and frequency usually refer to the same thing (higher order is equivalent to higher frequency), as $\cos^n(x) = a \cos(nx) + ...$. Therefore, I don't see the difference between "order compactness" and "frequency compactness." Regarding spatial compactness, the paper states: "Thanks to the Gaussian window, we can represent the signal or image in a well-localized spatial domain." Does "Gaussian window" refer to the hyperparameter $s_0$ in equation 1, which essentially controls the coefficients of $\cos^n(x)$ or $\cos(nx)$ in the Taylor expansion? If so, I believe the new terminology might make things more complicated. Overall, I recommend not introducing new terminology unless absolutely necessary. Relation to FourierNet: This framework seems very relevant to FourierNet and the paper "Fourier features let networks learn high frequency functions in low dimensional domains." To my understanding, the FourierNet papers use random cosine and sine functions to map coordinates to high dimensions, and then apply ReLU MLP. In this paper, you consider a geometrically spaced (non-random) mapping coupled with cosine and sine functions, and then apply MLP with activation functions $\exp(-x^2)$. Is this correct? Could you clarify the relationship and differences between these approaches in the paper? Since FourierNet is highly related, I would appreciate it if you could add FourierNet experiments as another baseline (if possible). Additionally, the FourierNet paper specifically mentions "learning high frequency functions in low dimensional domains," which is also a key novelty advocated in this paper. Could you elaborate on what you mean by "frequency limitation" in the table? Minor comments: In the first paragraph of the introduction, please add a reference for "Implicit Neural Representations" [citation!]. In Table 1, please add references for each network in the first column. Additionally, could you clarify what you mean by "principle" in the second column? Combining "principle" and "disadvantage" seems a bit unusual. The numbers reported in Figure 3 appear inconsistent with those reported in Figure 7 of the WIRE paper ("Wavelet Implicit Neural Representations"). Please clarify this discrepancy. Thanks! ———————————————————————————————————————————————————— ➡️ **So let's consider an image INR on a unit square $\Phi : [0,1] \times [0,1] \to \mathbb{R}^3$. If I understand correctly, using one single parameterization of the $W,b,\sigma$ (btw I think $\sigma$ the frequency right?), this function $\Phi$ can only represent one image right? So when you train the weights, do you only train one set of weights on a single image?** We thank to the reviewer for the comment. You are correct that the function $\Phi: [0,1] \times [0,1] \to \mathbb{R}^3$ represents a single image, with each pixel color (RGB) encoded by the function's output. When using a parameterisation of $W$, $b$, and $\sigma$ (where $\sigma$ represents frequency), this set of parameters will only encode one image. Training is conducted such that a single set of weights and associated parameters is optimised per image. Thus, each set is specific to the particular image on which it was trained. ➡️ **Related to the previous question, yes I now understand a bit better where the data comes from. So what is the pseudo algorithm for training? I might be missing something, but I genuinely do not see how I should be implementing the algorithm you are proposing in this paper.** We thank to the reviewer for the comment. We provide the pseudocode next. ____________________________________________________________ Pseudocode __________________________________________________________ Input: $\mathbf{x}$, the coordinate of the input image $y$ Output: Trained model $\Phi$ that can map coordinates $\mathbf{x}$ to pixel values $y$. Initialise network parameters $\mathbf{W}, \mathbf{b}, \sigma$ $\textbf{for}$ $i$ from $0$ to epoch_number $\textbf{do}$: &nbsp; &nbsp; &nbsp; #calculate the loss for the example denoising task &nbsp; $\qquad L(\Phi(\mathbf{x}), y)= \int_\Omega |\Phi(\mathbf{x})- y|^2 = \frac{1}{N}\sum_{j=1}^N|\Phi(\mathbf{x}_j)- y_j|^2$ &nbsp; &nbsp; &nbsp; Update the gradients to the network parameters $\textbf{end}$ ____________________________________________________________ ------------------------ ## Reviewer [hF2P] ➡️ **So let's consider an image INR on a unit square $\Phi : [0,1] \times [0,1] \to \mathbb{R}^3$. If I understand correctly, using one single parameterization of the $W,b,\sigma$ (btw I think $\sigma$ the frequency right?), this function $\Phi$ can only represent one image right? So when you train the weights, do you only train one set of weights on a single image?** We thank to the reviewer for the comment. You are correct that the function $\Phi: [0,1] \times [0,1] \to \mathbb{R}^3$ represents a single image, with each pixel color (RGB) encoded by the function's output. When using a parameterisation of $W$, $b$, and $\sigma$ (where $\sigma$ represents frequency), this set of parameters will only encode one image. Training is conducted such that a single set of weights and associated parameters is optimised per image. Thus, each set is specific to the particular image on which it was trained. ➡️ **Related to the previous question, yes I now understand a bit better where the data comes from. So what is the pseudo algorithm for training? I might be missing something, but I genuinely do not see how I should be implementing the algorithm you are proposing in this paper.** We thank to the reviewer for the comment. We provide the pseudocode next. \____________________________________________________________ Pseudocode \____________________________________________________________ Input: $\mathbf{x}$, the coordinate of the input image $y$ Output: Trained model $\Phi$ that can map coordinates $\mathbf{x}$ to pixel values $y$. Initialise network parameters $\mathbf{W}, \mathbf{b}, \sigma$ $\textbf{for}$ $i$ from $0$ to epoch_number $\textbf{do}$: #calculate the loss for the example denoising task $\qquad L(\Phi(\mathbf{x}), y) = \int_\Omega |\Phi(\mathbf{x})- y|^2 = \frac{1}{N}\sum_{j=1}^N|\Phi(\mathbf{x}_j)- y_j|^2$ Update the gradients to the network parameters $\textbf{end}$ \____________________________________________________________ ``` ## Reviewer [wgCG] ➡️ **I would clarify again, that by framing the approach in the current state-of-the-art I rather meant approaches different from INR. I understand that it is a novel field, so comparison within the field will not help to grasp the difference between previously used representations from deep networks. Moreover, the clarity of the algorithm can be also helpful - how exactly DNN is employed in such approaches.** Thank you for your question. We appreciate the opportunity to clarify. Implicit Neural Representations (INRs) inherently differ from traditional Deep Neural Networks (DNNs) by focusing on learning from a single data instance, rather than relying on extensive training datasets. This distinction underpins the unique capability of INRs to model continuous, resolution-independent representations directly from individual instances. Consequently, comparing INRs directly with DNNs, which are designed and optimised for learning from large datasets, may not provide a fair or meaningful assessment of their respective capabilities and advantages. Our approach aligns with the protocol in the INR domain, emphasising the specialised application of INRs and the different philosophical foundation compared to traditional DNNs. As INRs are solely based on MLPs (also known as coordinate MLPs). We clarify this distinction in the related work in blue colour. <!-- The implicit neural representation (INR) was implemented on Multilayer Perceptron (MLP) with our proposed activation function. We fed input coordinates representing points into the MLP which parameterises a function to predict scalar values for each input coordinate. These scalar values represent the implicit function values or properties associated with the input points. During training, the predicted values are compared to ground truth values using a loss function, which measures the discrepancy between the predicted and ground truth values. The MLP is then optimised to minimise this loss, adjusting its weights and biases through backpropagation. Once trained, the MLP can be used to predict implicit function values for new input coordinates, allowing for the reconstruction of implicit surfaces or other tasks such as surface rendering or mesh generation. --> ## Reviewer [o7eH] ➡️ **Continuing the question about the log-linear approach, it is confusing that the method used to determine the parameter mentioned essentially inside the sentence about the formula derivation. I suggest rephrasing the sentence before the last equation on page 5 or adding a sentence or two explaining it clearly in the paper.** We thank the reviewer for their suggestion. To address this comment, we have updated the manuscript to clarify the introduction of a frequency parameter $\sigma$ and the application of a log-linear strategy to enhance our representation's efficiency. This strategy is designed to balance magnitudes of different scales and aims to improve the output. ➡️ **I found the addition of Table 1 helpful. However, it now has some content that needs to be clarified. Specifically, why is the Wavelet transform an advantage and the ReLU function a disadvantage? I suggest listing the consequences of using these approaches instead. For instance, as I understand, instead of ReLU activation, the disadvantage of the FourierNet should be poor representation in the high-frequency domain. A more detailed explanation, for example, clarifying that it happens due to the ReLU activation, can be added to the table caption or in the text if needed.** We thank the reviewer for their insightful comment. We concur that the label "Advantage" in Table 1 may not adequately convey the intended message. Thus, we have revised this to "Principle" to more accurately reflect the core characteristics of each approach. Pertaining to the disadvantage associated with FourierNet, we have refined the description to "Frequency Limitation" to encapsulate the well-known challenges related to ReLU activation's handling of high-frequency components. All changes can be found in blue colour. ➡️ **In equation (5), the RHS of the first and second lines in black can be removed. The RHS of the first line is the same as the first line in blue, and the second line is not used.** We thank the reviewer for the suggestion. In the updated version, we have removed the suggeted lines. ➡️ **The last sentence of the conclusion is hard to parse because of several typos: "While there are still unexplored inverse problems pushing the boundaries of INR, our solution remains at the forefront of the already developed tasks using INR, but would be promising next step to challenge on the other inverse problem that yet to be explore in the community." I suggest rephrasing it to improve clarity. For example, something like "While there are still unexplored inverse problems pushing the boundaries of INR, our solution remains at the forefront of the already-developed tasks using INR. Nonetheless, it would be a promising next step to challenge the other inverse problems yet to be explored in the community.".** We thank the reviewer for their feedback. We agree that the readability of this part was challenging. We have made updates to clarify our message. ## Reviewer [hF2P] ➡️ **I still don't fully understand what you are referring to as an implicit neural representation. Is the point that you want to represent a dataset using a continuous function that interpolates the dataset? What does the output here mean? Can you give an example of what this representation is doing with respect to a dataset? What would the grid-based approach do for this example?** Thanks for the question. We appreciate the opportunity to further clarify the concept and application of Implicit Neural Representations (INR) in our work. It's important to note that the concept of INR is not novel to our study but is a well-established area of research within the literature. Our goal is to contribute with a new solution within the widely studied area of INR. We would like to address your questions separatly next, with updates in related work in blue colour. **[What we refer to INR?]** Firstly, at its core, an INR employs a neural network to model a continuous function that can interpolate or approximate the underlying data distribution of a dataset. This continuous function is learned from the discrete samples of the dataset, enabling the representation of complex data structures or spaces in a compact and continuous manner. **[Meaning of the output]** Secondly, the output of an INR typically represents the value of the continuous function at a given point in the input space. This can vary based on the application, such as generating the colour value at a specific pixel for image synthesis or the density at a point in space for 3D volume rendering. **[Example INR]** Thirdly, imagine we want to use an INR to model a 2D image. The network would learn to predict the colour of any given pixel (x, y) on data. For a given input coordinate, the network outputs the colour value, effectively allowing us to reconstruct the image or even generate new parts of the image by querying points not explicitly represented in the original dataset. **[Grid-based example]** Unlike grid-based methods that discretely represent data and are limited by resolution and dimensionality, INRs provide a continuous, resolution-independent representation. For the same 2D image example, a grid-based approach might store colour values at fixed pixel locations, leading to potential issues with scaling or transformation. INRs, however, can theoretically generate any pixel value within the image domain, offering a more flexible and scalable solution. ➡️ **Is $\Phi(x)$ what you are defining as the implicit neural representation? How are the parameters $\sigma, W, b$ chosen for your applications?** Yes, $\Phi(x)$ defines as the function modelled by the neural network to represent continuously across a domain. The parameters $\sigma, W, b$ —which represent the activation function, weights, and biases of the network, respectively—are crucial for defining the function's shape and complexity. Training the network involves adjusting $W , b$ to minimise the difference between the network's output and the actual data. We have added a clarified note on this. ➡️ **What's the significance of order compactness and frequency compactness? My point here being that yes this is a property, but why do you need this property and why would it help? In particular, I don't think the Stone--Weierstrass Theorem is a convincing reason, since Fourier series is can also approximate arbitrary functions in a compact interval.** Thanks for the question. We now clarify that order compactness and frequency compactness relate to how efficiently a model can approximate complex functions within a specific domain. While Fourier series can indeed approximate arbitrary functions, the compactness properties enhance the efficiency and effectiveness of learning by reducing the model's complexity and improving generalisation. Essentially, these properties guide the selection of functions that the model can approximate well, focusing on those that are significant for the application. This focus helps in achieving higher fidelity, making the learning process more efficient and leading to better performance. We have added a clarified note in blue colour. ➡️ **For Figure 1, I was hoping to get a full description of the experiment setup. E.g. what is the dataset here? What do you mean by the first 100 projections and how did you compute it? How did you fit the model? What exactly is the output and what does it mean? This is important because even a non-expert should be able to replicate your experiment from reading the paper.** We thank the reviewer for the comment. We would like to clarify that in the context of CT reconstruction, the term "projections" refers to the images obtained by passing X-rays through a patient from various angles. These projections are then used to reconstruct the patient's internal structures. CT projections are computed using a process called Radon transform. The data used is publicly available where the full acquisition is provided and we chose a subset of 100 projections following the standard protocol for this dataset. The input is a set of projects, as the main goal is to recontruct a high fidelity with less data, and the output is the reconstructed image. For sake of clarity, we pointed out the full description, of the setting of this visualisation, in the experimental results and add a reference about the basics of CT reconstruction in blue colour. ------------------------- ## Reviewer [hF2P] ➡️ **How is an implicit neural representation defined, and what is the purpose? I hope the authors can provide some additional context for readers who are not experts. At the same time, what is a traditional grid based approach?** We thank the reviewer for the comments. We would like to start by clarifying the concept of Implicit Neural Representations (INR), which serve as a modern method for encoding data structures or continuous signals using neural networks. Unlike traditional grid-based approaches that store information in discrete intervals, INRs represent data in a continuous, memory-efficient form, enabling more control and manipulation of the represented content. Secondly, we now clarify that a traditional grid-based approach, on the other hand, represents data on a fixed, discrete grid, like pixels for images or voxels for 3D shapes. This can be less efficient and more memory-intensive, especially for high-resolution data or complex geometries. That is why, INR can be used a wide range of applications. To address this comment, we have updated the text, and the changes can be found in the first paragraph of Section 1 in blue colour. ➡️ **Can you help me define the notations used in section 3? For example, what is the variable x, and what are the dimensions, what do the dots in equation 1 represent, and what is the output dimension of this function? What is the difference between x and bold x?** Thanks for the comment. We now clarify that the variable defines the input signal to each neuron in the layer, and the dimension of $x$ is the dimension of the input data according to different tasks. The dots in equation 1 represent the terms with different choices of $j \in \{1,2,...,m\}$. The $x$ in the later part of the methodology denotes general input into a function, while the bold $\mathbf{x}$ here in equation 1 and 2 denotes the data with the same dimension as input data. The $x$ in equation 1 should be the bold $\mathbf{x}$, we thank the reviewer for pointing out the typo. We have added a clarifying note to make this passage clearer. ➡️ **You introduced several concepts such as "order compactness" and "frequency compactness", but how are these terms defined precisely? It's unclear for me what these terms mean.** We thank to the reviewer for the comment. To addres this comment, we provide detailed context on these terms. Firstly, we refer to order compactness to the ability of a representation to encapsulate information efficiently across different orders of derivative or variation within the data. Moreover, 'order' typically relates to the smoothness and the level of detail that the representation can capture. . Mathematically, this can be associated with the representation's capacity to approximate functions and their derivatives to a certain degree of accuracy. Secondly, frequeny compatness refer to describe the efficiency with which a representation can encapsulate information at different frequencies. A representation with high frequency compactness can accurately reproduce both the overarching structures and the intricate details of the data, making it particularly valuable for applications requiring high fidelity. To add clarity, we have updated the manuscript to include an explicit discussion on these concepts. The changes can be seen in Section 3 in blue colour. ➡️ **Towards the end of section 3, you demonstrated an example of using INR in Figure 1. I'm not sure what you did here, as in what were the input and output of this experiment?** Thanks for the comment, we provide next details on the setting for Figure 1. The purpose of this figure is to illustrate the effect of our trilogy. The experiment is on CT reconstruction task using 100 projections to support our nonlinear trilogy: TRIDENT (trilogy), and TRIDENT without Frequency, Order, and Spatial Compactness separately. To address this comment, we have done the following changes. Firstly, we have added clarity in the setting used for the CT experiments. Secondly, we have reworded the description on the last paragraph of subsection 3.2 for clarity purposes. All changes can be seen in blue colour. ➡️ **One small thing: I do want to briefly discuss the section headings with the finger pointing emoji. I'm not sure what the TMLR formatting requirements are, but I found these emoji distracting and unprofessional. Regardless of the formatting requirements, I would recommend you avoid using these, so the readers can focus on the content.** We acknowledge that style preferences can vary significantly among readers and reviewers. It's worth noting that other reviewers have not commented on this aspect, which might suggest a range of opinions on such stylistic choices. However, we deeply respect the scholarly standards and the reading experience of all reviewers and readers. If you have strong feelings on this matter, we are more than willing to reconsider our stylistic choices to ensure that our manuscript. ## Reviewer [wgCG] ➡️ **Please add the explanation of the approach in general, frame it in the state of the art reconstruction approaches. Maybe even evaluation compared to them would be interesting (if applicable).** Thank you for your valuable feedback and suggestion to elaborate on our approach within the context of state-of-the-art reconstruction methodologies. We agree that a comprehensive explanation and framing within the current landscape would enhance the manuscript and appreciate the opportunity to clarify how our approach distinguishes itself. In our work, we introduce a new approach that is characterised by a unique mathematical components: frequency compactness, spatial compactness, and order compactness. To the best of our knowledge, this triad has not been collectively considered in the context of implicit neural representations. Frequency compactness allows our model to effectively represent and reconstruct signals at various scales, improving the fidelity of high-frequency details. Spatial compactness ensures that our representations are localized and precise, facilitating the accurate rendering of intricate spatial variations. Lastly, order compactness contributes to the smoothness and continuity of the represented data, which is particularly beneficial for complex geometries and textures. In our current work, we compared our TRIDENT technique against the state-of-the-art techniques. However, to provide more clarity, we have done the following changes. Firstly, we added a clarifying notes on the conceptual benefit of our trilogy in Section 3 (see changes in blue colour). Secondly, we have added a new table displaying the advantages and disadvatanges of existing INRs techniques and ours. ➡️ **I would argue that showing image of the sound wave is not efficient for understanding the results (fig.7 and fig.10). Maybe just providing link to the audiofiles is better.** We thank the reviewer to point a great option for efficiently interpreting the results, we have attached the folder links here with all the audio files. For the audio reconstruction in Fig.7: https://drive.google.com/drive/folders/1i7t3_z9FhJu8hbBZFQOJJESWBgj9LxgT?usp=sharing. For the audio reconstruction in Fig.10: https://drive.google.com/drive/folders/113YPlovkwA14WGu3o-_K8rOJe-KEErjz?usp=drive_link We will include the audio results directly in our project website, which will be available upon acceptance due to anonymous reason. ➡️ **Ablation study shows that the approach is very sensitive to some hyperparameters, but no hints on how to adjust them are not given.** Thanks for the comment. We adhere to a standard protocol by employing a grid search technique to identify the optimal hyperparameter values for our approach. This methodical search allows us to systematically explore a range of possible values and evaluate their impact on the performance of our model. By doing so, we ensure that the chosen hyperparameters are well-suited to maximise the efficacy of our representation method. We therefore provide a set of recommended hyperparameter values that we found to yield the best results in our experiments. To address this comment, we have added a clarifying note in Subsection 4.1 in blue colour. ➡️ **Please clarify what are j and m stand for in formula 1 and also what should be there in the ... parts.** $m$ stands for the mapping size and $j$ stands for the mapping index, ranging from $1$ to $m$, while the dots in equation 1 represent the terms with different choices of $j$. We clarify all the notations in the updated manuscript in blue. ➡️ **Typo in the reference to Kingma et al. in section 4.1** We thank the reviewer for pointing out the typo, the change has been made in our updated manuscript in blue. ➡️ **In equation 5 the summation index switches from n to i** The summation index was changed due to a change of variable during derivation. To clarify the process, we add additional derivation steps in equation 5 in blue in the updated manuscript. ➡️ **It seems that there is a typo in the last sentence of section 3.1, since \phi_i is not taking part in definition of \gamma** We thanks the reviewer for pointing out the typo, it should be ‘and’ not ‘where’, the correction has been made in blue in the updated version. ➡️ **Missing reference in the third paragraph of the introduction ([ref])** We thanks the reviewer for pointing out the typo, the additional reference had been made in blue in the updated version. ## Reviewer [o7eH] ➡️ **Add a more in-depth summary of the work and discussion of the method instead of the statement "TRIDENT achieves always the best and hence it is the best solution for nonlinear INRs." appearing in the conclusion. It would be particularly valuable to see a discussion of the method's potential limitations and possible further work.** We thank the reviewer for their valuable feedback. In response, we have updated our manuscript to better articulate the advantages of TRIDENT, emphasising its integration of order compactness, frequency compactness, and spatial compactness for superior performance in nonlinear INRs. Additionally, we've introduced a discussion on the method's limitations and potential directions for future research. The changes can be found in blue colour. ➡️ **Clarifying the theoretical explanation** * **What is the derivation for the equation in the second line of equation (5)?** We are happy to provide detailed derivation of \begin{equation} \begin{aligned} \sum_{n=0}^{\infty} \frac{(-1)^{n}}{2^{n} n !} \cos ^{n}(2 x)+\frac{(-1)^{n+1}}{2^{n+1} n !} \cos ^{n}(2 x) %+\frac{(-1)^{n+2}}{2^{n+2} n ! 2!} \cos ^{n}(2 x) +\frac{(-1)^{n+3}}{2^{n+3} n ! 3 !} \cos ^{n}(2 x) +\cdots = \sum_{i=0}^{\infty} \frac{(-1)^i}{2^{i} i !} \left[\cos^{i}(2x)+\binom{i}{1}\cos^{i-1}(2x)+\cdots\right] \end{aligned} \end{equation} \begin{array}{l} LHS =\sum_{n=0}^{\infty}\left(\sum_{j=0}^{\infty} \frac{(-1)^{n+j}}{2^{n+j} \cdot(n !) \cdot(j !)}\right) \cos ^{n}(2 x) \\ =\sum_{n=0}^{\infty}\left(\sum_{j=0}^{\infty} \frac{(-1)^{n+j}}{2^{n+j}(n+j) !} \frac{(n+j) !}{n ! j !} \cos ^{n}(2 x)\right) \end{array} Let $i=n+j$ , then \begin{array}{l} =\sum_{i=0}^{\infty}\left( \sum_{j=0}^{i}\frac{{i!}}{(i-j)! j!} \cos ^{j}(2 x)\right) \cdot \frac{(-1)^{i}}{2^{i} i !} \\ =\sum_{i=0}^{\infty}\left(\cos ^{i}(2 x)+ \binom{i}{1} \cos ^{i-1}(2 x)+\cdots\right) \frac{(-1)^{i}}{2^{i} i !} \\ =RHS \end{array} * **What exactly is meant by the "log-linear" strategy at the end of Section 3, and how is it used to get the final function?** The 'log-linear' strategy at the end of section 3 is introduced to adjust the frequency parameter $\sigma$ that we introduced to get the final function. We trained the model by adjusting the parameter $\sigma$ to maximise the likelihood of the training data with Adam optimisation algorithm. * **In section 3.1, I suggest explaining what j, W, and b are and clearly stating which parameters are trained and which are chosen in advance.** The parameter $j$ is the mapping index with range $1$ to $m$, while $W$ and $b$ are the parameters used in the neural network, as weight and bias. We thanks the reviewer to pointing out the confusion, and updated a clarification in blue in the updated manuascript. * **As I understand, the first x appearing in (1) should be in bold.** The $x$ in equation 1 should indeed be in bold as $\mathbf{x}$, we thank the reviewer for pointing out the typo. The modification has been made in blue in the updated document. * **I suggest replacing "without losing generality" with "without loss of generality.", which is the usual expression.** We thank the reviewer for the suggestion, it is updated accordingly in blue. ➡️ **Minor (Not critical to securing my recommendation) Improving the writing** Thanks for the insighful comments. We have updated the manuscript addressing every single minor suggestion. All changes can be found in blue colour.

    Import from clipboard

    Paste your markdown or webpage here...

    Advanced permission required

    Your current role can only read. Ask the system administrator to acquire write and comment permission.

    This team is disabled

    Sorry, this team is disabled. You can't edit this note.

    This note is locked

    Sorry, only owner can edit this note.

    Reach the limit

    Sorry, you've reached the max length this note can be.
    Please reduce the content or divide it to more notes, thank you!

    Import from Gist

    Import from Snippet

    or

    Export to Snippet

    Are you sure?

    Do you really want to delete this note?
    All users will lose their connection.

    Create a note from template

    Create a note from template

    Oops...
    This template has been removed or transferred.
    Upgrade
    All
    • All
    • Team
    No template.

    Create a template

    Upgrade

    Delete template

    Do you really want to delete this template?
    Turn this template into a regular note and keep its content, versions, and comments.

    This page need refresh

    You have an incompatible client version.
    Refresh to update.
    New version available!
    See releases notes here
    Refresh to enjoy new features.
    Your user state has changed.
    Refresh to load new user state.

    Sign in

    Forgot password

    or

    By clicking below, you agree to our terms of service.

    Sign in via Facebook Sign in via Twitter Sign in via GitHub Sign in via Dropbox Sign in with Wallet
    Wallet ( )
    Connect another wallet

    New to HackMD? Sign up

    Help

    • English
    • 中文
    • Français
    • Deutsch
    • 日本語
    • Español
    • Català
    • Ελληνικά
    • Português
    • italiano
    • Türkçe
    • Русский
    • Nederlands
    • hrvatski jezik
    • język polski
    • Українська
    • हिन्दी
    • svenska
    • Esperanto
    • dansk

    Documents

    Help & Tutorial

    How to use Book mode

    Slide Example

    API Docs

    Edit in VSCode

    Install browser extension

    Contacts

    Feedback

    Discord

    Send us email

    Resources

    Releases

    Pricing

    Blog

    Policy

    Terms

    Privacy

    Cheatsheet

    Syntax Example Reference
    # Header Header 基本排版
    - Unordered List
    • Unordered List
    1. Ordered List
    1. Ordered List
    - [ ] Todo List
    • Todo List
    > Blockquote
    Blockquote
    **Bold font** Bold font
    *Italics font* Italics font
    ~~Strikethrough~~ Strikethrough
    19^th^ 19th
    H~2~O H2O
    ++Inserted text++ Inserted text
    ==Marked text== Marked text
    [link text](https:// "title") Link
    ![image alt](https:// "title") Image
    `Code` Code 在筆記中貼入程式碼
    ```javascript
    var i = 0;
    ```
    var i = 0;
    :smile: :smile: Emoji list
    {%youtube youtube_id %} Externals
    $L^aT_eX$ LaTeX
    :::info
    This is a alert area.
    :::

    This is a alert area.

    Versions and GitHub Sync
    Get Full History Access

    • Edit version name
    • Delete

    revision author avatar     named on  

    More Less

    Note content is identical to the latest version.
    Compare
      Choose a version
      No search result
      Version not found
    Sign in to link this note to GitHub
    Learn more
    This note is not linked with GitHub
     

    Feedback

    Submission failed, please try again

    Thanks for your support.

    On a scale of 0-10, how likely is it that you would recommend HackMD to your friends, family or business associates?

    Please give us some advice and help us improve HackMD.

     

    Thanks for your feedback

    Remove version name

    Do you want to remove this version name and description?

    Transfer ownership

    Transfer to
      Warning: is a public team. If you transfer note to this team, everyone on the web can find and read this note.

        Link with GitHub

        Please authorize HackMD on GitHub
        • Please sign in to GitHub and install the HackMD app on your GitHub repo.
        • HackMD links with GitHub through a GitHub App. You can choose which repo to install our App.
        Learn more  Sign in to GitHub

        Push the note to GitHub Push to GitHub Pull a file from GitHub

          Authorize again
         

        Choose which file to push to

        Select repo
        Refresh Authorize more repos
        Select branch
        Select file
        Select branch
        Choose version(s) to push
        • Save a new version and push
        • Choose from existing versions
        Include title and tags
        Available push count

        Pull from GitHub

         
        File from GitHub
        File from HackMD

        GitHub Link Settings

        File linked

        Linked by
        File path
        Last synced branch
        Available push count

        Danger Zone

        Unlink
        You will no longer receive notification when GitHub file changes after unlink.

        Syncing

        Push failed

        Push successfully