Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 14 de 14
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
1.
Clin Nucl Med ; 49(8): 793-796, 2024 Aug 01.
Artigo em Inglês | MEDLINE | ID: mdl-38886924

RESUMO

ABSTRACT: The flare phenomenon is a transient increase in the number or intensity of lesions on bone scans after treatment, signifying curative effect. DOTA-ibandronic acid (DOTA-IBA) is a new prodrug that targets bone metastases and can be labeled with 177 Lu. Here, we report the case of a 58-year-old woman with bone metastasis, in whom the flare phenomenon was observed after 4 cycles of 177 Lu-DOTA-IBA treatment. No adverse effects were observed during the treatment and follow-up periods.


Assuntos
Neoplasias Ósseas , Neoplasias Pulmonares , Humanos , Feminino , Pessoa de Meia-Idade , Neoplasias Ósseas/secundário , Neoplasias Ósseas/diagnóstico por imagem , Neoplasias Ósseas/radioterapia , Neoplasias Pulmonares/diagnóstico por imagem , Neoplasias Pulmonares/patologia , Neoplasias Pulmonares/radioterapia , Lutécio/efeitos adversos , Compostos Organometálicos , Radioisótopos
2.
Artigo em Inglês | MEDLINE | ID: mdl-38656847

RESUMO

This article aims to solve the video object segmentation (VOS) task in a scribble-supervised manner, in which VOS models are not only initialized with sparse target scribbles for inference but also trained by sparse scribble annotations. Thus, the annotation burdens for both initialization and training can be substantially lightened. The difficulties of scribble-supervised VOS lie in two aspects: 1) it demands a strong reasoning ability to carefully segment the target given only a sparse initial target scribble and 2) it necessitates learning dense prediction from sparse scribble annotations during training, requiring powerful learning capability. In this work, we propose a reliability-guided hierarchical memory network (RHMNet) for this task, which segments the target in a stepwise expanding strategy w.r.t. the memory reliability level. To be specific, RHMNet maintains a reliability-guided memory bank. It first uses the high-reliability memory to locate the region with high reliability belonging to the target, i.e., highly similar to the initial target scribble. Then, it expands the located high-reliability region to the entire target conditioned on the region itself and all existing memories. In addition, we propose a scribble-supervised learning mechanism to facilitate the model learning for dense prediction. It exploits the pixel-level relations within a single frame and the instance-level variations across multiple frames to take full advantage of the scribble annotations in sequence training samples. The favorable performance on four popular benchmarks demonstrates that our method is promising. Our project is available at: https://github.com/mkg1204/RHMNet-for-SSVOS.

3.
EJNMMI Res ; 14(1): 30, 2024 Mar 22.
Artigo em Inglês | MEDLINE | ID: mdl-38517637

RESUMO

BACKGROUND: We designed and synthesized a novel bisphosphonate radiopharmaceutical (68 Ga- or 177Lu-labeled DOTA-ibandronate [68 Ga/177Lu-DOTA-IBA]) for the targeted diagnosis and treatment of bone metastases. The biodistribution and internal dosimetry of a single therapeutic dose of 177Lu-DOTA-IBA were evaluated using a series of single-photon emission computerized tomography (SPECT) images and blood samples. Five patients with multiple bone metastases were included in this prospective study. After receiving 1110 MBq 177Lu-DOTA-IBA, patients underwent whole-body planar, SPECT/CT imaging and venous blood sampling over 7 days. Dosimetric evaluation was performed for the main organs and tumor lesions. Safety was assessed using blood biomarkers. RESULTS: 177Lu-DOTA-IBA showed fast uptake, high retention in bone lesions, and rapid clearance from the bloodstream in all patients. In this cohort, the average absorbed doses (ADs) in the bone tumor lesions, kidneys, liver, spleen, red marrow, bladder-wall, and osteogenic cells were 5.740, 0.114, 0.095, 0.121, 0.095, and 0.333 Gy/GBq, respectively. Although no patient reached the predetermined dose thresholds, the red marrow will be the dose-limiting organ. There were no adverse reactions recorded after the administration of 1110 MBq 177Lu-DOTA-IBA. CONCLUSION: Dosimetric results show that the ADs for critical organs and total body are within the safety limit and with high bone retention. It is a promising radiopharmaceutical alternative for the targeted treatment of bone metastases, controlling its progression, and improving the survival and quality of life of patients with advanced bone metastasis.

4.
Artigo em Inglês | MEDLINE | ID: mdl-38393839

RESUMO

Few-shot classification aims to adapt classifiers trained on base classes to novel classes with a few shots. However, the limited amount of training data is often inadequate to represent the intraclass variations in novel classes. This can result in biased estimation of the feature distribution, which in turn results in inaccurate decision boundaries, especially when the support data are outliers. To address this issue, we propose a feature enhancement method called CORrelation-guided feature Enrichment that generates improved features for novel classes using weak supervision from the base classes. The proposed CORrelation-guided feature Enhancement (CORE) method utilizes an autoencoder (AE) architecture but incorporates classification information into its latent space. This design allows the CORE to generate more discriminative features while discarding irrelevant content information. After being trained on base classes, CORE's generative ability can be transferred to novel classes that are similar to those in the base classes. By using these generative features, we can reduce the estimation bias of the class distribution, which makes few-shot learning (FSL) less sensitive to the selection of support data. Our method is generic and flexible and can be used with any feature extractor and classifier. It can be easily integrated into existing FSL approaches. Experiments with different backbones and classifiers show that our proposed method consistently outperforms existing methods on various widely used benchmarks.

5.
Front Mol Biosci ; 10: 1210347, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37780209

RESUMO

Theranostic in nuclear medicine combines diagnostic imaging and internal irradiation therapy using different therapeutic nuclear probes for visual diagnosis and precise treatment. GLP-1R is a popular receptor target in endocrine diseases, non-alcoholic steatohepatitis, tumors, and other areas. Likewise, it has also made breakthroughs in the development of molecular imaging. It was recognized that GLP-1R imaging originated from the study of insulinoma and afterwards was expanded in application including islet transplantation, pancreatic ß-cell mass measurement, and ATP-dependent potassium channel-related endocrine diseases. Fortunately, GLP-1R molecular imaging has been involved in ischemic cardiomyocytes and neurodegenerative diseases. These signs illustrate the power of GLP-1R molecular imaging in the development of medicine. However, it is still limited to imaging diagnosis research in the current molecular imaging environment. The lack of molecular-targeted therapeutics related report hinders its radiology theranostic. In this article, the current research status, challenges, and emerging opportunities for GLP-1R molecular imaging are discussed in order to open a new path for theranostics and to promote the evolution of molecular medicine.

7.
Artigo em Inglês | MEDLINE | ID: mdl-37195841

RESUMO

People may perform diverse gestures affected by various mental and physical factors when speaking the same sentences. This inherent one-to-many relationship makes co-speech gesture generation from audio particularly challenging. Conventional CNNs/RNNs assume one-to-one mapping, and thus tend to predict the average of all possible target motions, easily resulting in plain/boring motions during inference. So we propose to explicitly model the one-to-many audio-to-motion mapping by splitting the cross-modal latent code into shared code and motion-specific code. The shared code is expected to be responsible for the motion component that is more correlated to the audio while the motion-specific code is expected to capture diverse motion information that is more independent of the audio. However, splitting the latent code into two parts poses extra training difficulties. Several crucial training losses/strategies, including relaxed motion loss, bicycle constraint, and diversity loss, are designed to better train the VAE. Experiments on both 3D and 2D motion datasets verify that our method generates more realistic and diverse motions than previous state-of-the-art methods, quantitatively and qualitatively. Besides, our formulation is compatible with discrete cosine transformation (DCT) modeling and other popular backbones (i.e. RNN, Transformer). As for motion losses and quantitative motion evaluation, we find structured losses/metrics (e.g. STFT) that consider temporal and/or spatial context complement the most commonly used point-wise losses (e.g. PCK), resulting in better motion dynamics and more nuanced motion details. Finally, we demonstrate that our method can be readily used to generate motion sequences with user-specified motion clips on the timeline.

8.
Artigo em Inglês | MEDLINE | ID: mdl-37018296

RESUMO

While deep-learning-based tracking methods have achieved substantial progress, they entail large-scale and high-quality annotated data for sufficient training. To eliminate expensive and exhaustive annotation, we study self-supervised (SS) learning for visual tracking. In this work, we develop the crop-transform-paste operation, which is able to synthesize sufficient training data by simulating various appearance variations during tracking, including appearance variations of objects and background interference. Since the target state is known in all synthesized data, existing deep trackers can be trained in routine ways using the synthesized data without human annotation. The proposed target-aware data-synthesis method adapts existing tracking approaches within a SS learning framework without algorithmic changes. Thus, the proposed SS learning mechanism can be seamlessly integrated into existing tracking frameworks to perform training. Extensive experiments show that our method: 1) achieves favorable performance against supervised (Su) learning schemes under the cases with limited annotations; 2) helps deal with various tracking challenges such as object deformation, occlusion (OCC), or background clutter (BC) due to its manipulability; 3) performs favorably against the state-of-the-art unsupervised tracking methods; and 4) boosts the performance of various state-of-the-art Su learning frameworks, including SiamRPN++, DiMP, and TransT.

9.
IEEE Trans Image Process ; 32: 2003-2016, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-35839180

RESUMO

Typical methods for pedestrian detection focus on either tackling mutual occlusions between crowded pedestrians, or dealing with the various scales of pedestrians. Detecting pedestrians with substantial appearance diversities such as different pedestrian silhouettes, different viewpoints or different dressing, remains a crucial challenge. Instead of learning each of these diverse pedestrian appearance features individually as most existing methods do, we propose to perform contrastive learning to guide the feature learning in such a way that the semantic distance between pedestrians with different appearances in the learned feature space is minimized to eliminate the appearance diversities, whilst the distance between pedestrians and background is maximized. To facilitate the efficiency and effectiveness of contrastive learning, we construct an exemplar dictionary with representative pedestrian appearances as prior knowledge to construct effective contrastive training pairs and thus guide contrastive learning. Besides, the constructed exemplar dictionary is further leveraged to evaluate the quality of pedestrian proposals during inference by measuring the semantic distance between the proposal and the exemplar dictionary. Extensive experiments on both daytime and nighttime pedestrian detection validate the effectiveness of the proposed method.

10.
Artigo em Inglês | MEDLINE | ID: mdl-36279339

RESUMO

Real-world data usually present long-tailed distributions. Training on imbalanced data tends to render neural networks perform well on head classes while much worse on tail classes. The severe sparseness of training instances for the tail classes is the main challenge, which results in biased distribution estimation during training. Plenty of efforts have been devoted to ameliorating the challenge, including data resampling and synthesizing new training instances for tail classes. However, no prior research has exploited the transferable knowledge from head classes to tail classes for calibrating the distribution of tail classes. In this article, we suppose that tail classes can be enriched by similar head classes and propose a novel distribution calibration (DC) approach named as label-aware DC (). transfers the statistics from relevant head classes to infer the distribution of tail classes. Sampling from calibrated distribution further facilitates rebalancing the classifier. Experiments on both image and text long-tailed datasets demonstrate that significantly outperforms existing methods. The visualization also shows that provides a more accurate distribution estimation.

11.
IEEE Trans Image Process ; 30: 4867-4882, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-33950841

RESUMO

Restoring the clean background from the superimposed images containing a noisy layer is the common crux of a classical category of tasks on image restoration such as image reflection removal, image deraining and image dehazing. These tasks are typically formulated and tackled individually due to diverse and complicated appearance patterns of noise layers within the image. In this work we present the Deep-Masking Generative Network (DMGN), which is a unified framework for background restoration from the superimposed images and is able to cope with different types of noise. Our proposed DMGN follows a coarse-to-fine generative process: a coarse background image and a noise image are first generated in parallel, then the noise image is further leveraged to refine the background image to achieve a higher-quality background image. In particular, we design the novel Residual Deep-Masking Cell as the core operating unit for our DMGN to enhance the effective information and suppress the negative information during image generation via learning a gating mask to control the information flow. By iteratively employing this Residual Deep-Masking Cell, our proposed DMGN is able to generate both high-quality background image and noisy image progressively. Furthermore, we propose a two-pronged strategy to effectively leverage the generated noise image as contrasting cues to facilitate the refinement of the background image. Extensive experiments across three typical tasks for image background restoration, including image reflection removal, image rain steak removal and image dehazing, show that our DMGN consistently outperforms state-of-the-art methods specifically designed for each single task.

12.
IEEE Trans Image Process ; 30: 907-920, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-33259297

RESUMO

Person re-identification aims to identify whether pairs of images belong to the same person or not. This problem is challenging due to large differences in camera views, lighting and background. One of the mainstream in learning CNN features is to design loss functions which reinforce both the class separation and intra-class compactness. In this paper, we propose a novel Orthogonal Center Learning method with Subspace Masking for person re-identification. We make the following contributions: 1) we develop a center learning module to learn the class centers by simultaneously reducing the intra-class differences and inter-class correlations by orthogonalization; 2) we introduce a subspace masking mechanism to enhance the generalization of the learned class centers; and 3) we propose to integrate the average pooling and max pooling in a regularizing manner that fully exploits their powers. Extensive experiments show that our proposed method consistently outperforms the state-of-the-art methods on large-scale ReID datasets including Market-1501, DukeMTMC-ReID, CUHK03 and MSMT17.

13.
Artigo em Inglês | MEDLINE | ID: mdl-31670671

RESUMO

The main challenges of age estimation from facial expression videos lie not only in the modeling of the static facial appearance, but also in the capturing of the temporal facial dynamics. Traditional techniques to this problem focus on constructing handcrafted features to explore the discriminative information contained in facial appearance and dynamics separately. This relies on sophisticated feature-refinement and framework-design. In this paper, we present an end-toend architecture for age estimation, called Spatially-Indexed Attention Model (SIAM), which is able to simultaneously learn both the appearance and dynamics of age from raw videos of facial expressions. Specifically, we employ convolutional neural networks to extract effective latent appearance representations and feed them into recurrent networks to model the temporal dynamics. More importantly, we propose to leverage attention models for salience detection in both the spatial domain for each single image and the temporal domain for the whole video as well. We design a specific spatially-indexed attention mechanism among the convolutional layers to extract the salient facial regions in each individual image, and a temporal attention layer to assign attention weights to each frame. This two-pronged approach not only improves the performance by allowing the model to focus on informative frames and facial areas, but it also offers an interpretable correspondence between the spatial facial regions as well as temporal frames, and the task of age estimation. We demonstrate the strong performance of our model in experiments on a large, gender-balanced database with 400 subjects with ages spanning from 8 to 76 years. Experiments reveal that our model exhibits significant superiority over the state-of-the-art methods given sufficient training data.

14.
IEEE Trans Neural Netw Learn Syst ; 29(4): 920-931, 2018 04.
Artigo em Inglês | MEDLINE | ID: mdl-28141534

RESUMO

We present a new model for multivariate time-series classification, called the hidden-unit logistic model (HULM), that uses binary stochastic hidden units to model latent structure in the data. The hidden units are connected in a chain structure that models temporal dependencies in the data. Compared with the prior models for time-series classification such as the hidden conditional random field, our model can model very complex decision boundaries, because the number of latent states grows exponentially with the number of hidden units. We demonstrate the strong performance of our model in experiments on a variety of (computer vision) tasks, including handwritten character recognition, speech recognition, facial expression, and action recognition. We also present a state-of-the-art system for facial action unit detection based on the HULM.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...