SETTING: An Asian cohort in 16 pediatric HIV services across 6 countries.
METHODS: From 2005 to 2014, patients younger than 20 years who achieved virologic suppression and had subsequent viral load testing were included. Early virologic failure was defined as a HIV RNA ≥1000 copies per milliliter within 12 months of virologic suppression, and late virologic as a HIV RNA ≥1000 copies per milliliter after 12 months following virologic suppression. Characteristics at combination antiretroviral therapy initiation and virologic suppression were described, and a competing risk time-to-event analysis was used to determine cumulative incidence of virologic failure and factors at virologic suppression associated with early and late virologic failure.
RESULTS: Of 1105 included in the analysis, 182 (17.9%) experienced virologic failure. The median age at virologic suppression was 6.9 years, and the median time to virologic failure was 24.6 months after virologic suppression. The incidence rate for a first virologic failure event was 3.3 per 100 person-years. Factors at virologic suppression associated with late virologic failure included older age, mostly rural clinic setting, tuberculosis, protease inhibitor-based regimens, and early virologic failure. No risk factors were identified for early virologic failure.
CONCLUSIONS: Around 1 in 5 experienced virologic failure in our cohort after achieving virologic suppression. Targeted interventions to manage complex treatment scenarios, including adolescents, tuberculosis coinfection, and those with poor virologic control are required.
METHODS: Regional Asian data (2001-2016) were analyzed to describe PHIVA who experienced ≥2 weeks of lamivudine or emtricitabine monotherapy or treatment interruption and trends in CD4 count and HIV viral load during and after episodes. Survival analyses were used for World Health Organization (WHO) stage III/IV clinical and immunologic event-free survival during monotherapy or treatment interruption, and a Poisson regression to determine factors associated with monotherapy or treatment interruption.
RESULTS: Of 3,448 PHIVA, 84 (2.4%) experienced 94 monotherapy episodes, and 147 (4.3%) experienced 174 treatment interruptions. Monotherapy was associated with older age, HIV RNA >400 copies/mL, younger age at ART initiation, and exposure to ≥2 combination ART regimens. Treatment interruption was associated with CD4 count <350 cells/μL, HIV RNA ≥1,000 copies/mL, ART adverse event, and commencing ART age ≥10 years compared with age <3 years. WHO clinical stage III/IV 1-year event-free survival was 96% and 85% for monotherapy and treatment interruption cohorts, respectively. WHO immunologic stage III/IV 1-year event-free survival was 52% for both cohorts. Those who experienced monotherapy or treatment interruption for more than 6 months had worse immunologic and virologic outcomes.
CONCLUSIONS: Until challenges of treatment adherence, engagement in care, and combination ART durability/tolerability are met, monotherapy and treatment interruption will lead to poor long-term outcomes.
METHODS: Data collected 2001 to 2016 from PHIVA 10-19 years of age within a regional Asian cohort were analyzed using competing risk time-to-event and Poisson regression analyses to describe the nature and incidence of morbidity events and hospitalizations and identify factors associated with disease-related, treatment-related and overall morbidity. Morbidity was defined according to World Health Organization clinical staging criteria and U.S. National Institutes of Health Division of AIDS criteria.
RESULTS: A total 3,448 PHIVA contributed 17,778 person-years. Median age at HIV diagnosis was 5.5 years, and ART initiation was 6.9 years. There were 2,562 morbidity events and 307 hospitalizations. Cumulative incidence for any morbidity was 51.7%, and hospitalization was 10.0%. Early adolescence was dominated by disease-related infectious morbidity, with a trend toward noninfectious and treatment-related morbidity in later adolescence. Higher overall morbidity rates were associated with a CD4 count <350 cells/µL, HIV viral load ≥10,000 copies/mL and experiencing prior morbidity at age <10 years. Lower overall morbidity rates were found for those 15-19 years of age compared with 10-14 years and those who initiated ART at age 5-9 years compared with <5 or ≥10 years.
CONCLUSIONS: Half of our PHIVA cohort experienced a morbidity event, with a trend from disease-related infectious events to treatment-related and noninfectious events as PHIVA age. ART initiation to prevent immune system damage, optimize virologic control and minimize childhood morbidity are key to limiting adolescent morbidity.
SETTING: Asian regional cohort incorporating 16 pediatric HIV services across 6 countries.
METHODS: Data from PHIVA (aged 10-19 years) who received combination antiretroviral therapy 2007-2016 were used to analyze LTFU through (1) an International epidemiology Databases to Evaluate AIDS (IeDEA) method that determined LTFU as >90 days late for an estimated next scheduled appointment without returning to care and (2) the absence of patient-level data for >365 days before the last data transfer from clinic sites. Descriptive analyses and competing-risk survival and regression analyses were used to evaluate LTFU epidemiology and associated factors when analyzed using each method.
RESULTS: Of 3509 included PHIVA, 275 (7.8%) met IeDEA and 149 (4.3%) met 365-day absence LTFU criteria. Cumulative incidence of LTFU was 19.9% and 11.8% using IeDEA and 365-day absence criteria, respectively. Risk factors for LTFU across both criteria included the following: age at combination antiretroviral therapy initiation <5 years compared with age ≥5 years, rural clinic settings compared with urban clinic settings, and high viral loads compared with undetectable viral loads. Age 10-14 years compared with age 15-19 years was another risk factor identified using 365-day absence criteria but not IeDEA LTFU criteria.
CONCLUSIONS: Between 12% and 20% of PHIVA were determined LTFU with treatment fatigue and rural treatment settings consistent risk factors. Better tracking of adolescents is required to provide a definitive understanding of LTFU and optimize evidence-based models of care.
METHODS: Data on children with perinatally acquired HIV aged <18 years on first-line, non-nucleoside reverse transcriptase inhibitor-based cART with viral suppression (two consecutive pVL <400 copies/mL over a six-month period) were included from a regional cohort study; those exposed to prior mono- or dual antiretroviral treatment were excluded. Frequency of pVL monitoring was determined at the site-level based on the median rate of pVL measurement: annual 0.75 to 1.5, and semi-annual >1.5 tests/patient/year. Treatment failure was defined as virologic failure (two consecutive pVL >1000 copies/mL), change of antiretroviral drug class, or death. Baseline was the date of the second consecutive pVL <400 copies/mL. Competing risk regression models were used to identify predictors of treatment failure.
RESULTS: During January 2008 to March 2015, there were 1220 eligible children from 10 sites that performed at least annual pVL monitoring, 1042 (85%) and 178 (15%) were from sites performing annual (n = 6) and semi-annual pVL monitoring (n = 4) respectively. Pre-cART, 675 children (55%) had World Health Organization clinical stage 3 or 4, the median nadir CD4 percentage was 9%, and the median pVL was 5.2 log10 copies/mL. At baseline, the median age was 9.2 years, 64% were on nevirapine-based regimens, the median cART duration was 1.6 years, and the median CD4 percentage was 26%. Over the follow-up period, 258 (25%) CLWH with annual and 40 (23%) with semi-annual pVL monitoring developed treatment failure, corresponding to incidence rates of 5.4 (95% CI: 4.8 to 6.1) and 4.3 (95% CI: 3.1 to 5.8) per 100 patient-years of follow-up respectively (p = 0.27). In multivariable analyses, the frequency of pVL monitoring was not associated with treatment failure (adjusted hazard ratio: 1.12; 95% CI: 0.80 to 1.59).
CONCLUSIONS: Annual compared to semi-annual pVL monitoring was not associated with an increased risk of treatment failure in our cohort of virally suppressed children with perinatally acquired HIV on first-line NNRTI-based cART.
METHODS: CLHIV aged <18 years, who were on first-line cART for ≥12 months, and had virological suppression (two consecutive plasma viral load [pVL] <50 copies/mL) were included. Those who started treatment with mono/dual antiretroviral therapy, had a history of treatment interruption >14 days, or received treatment and care at sites with a pVL lower limit of detection >50 copies/mL were excluded. LLV was defined as a pVL 50 to 1000 copies/mL, and VF as a single pVL >1000 copies/mL. Baseline was the time of the second pVL
METHODS: We used Cox regression to analyze data of a cohort of Asian children.
RESULTS: A total of 2608 children were included; median age at cART was 5.7 years. Time-updated weight for age z score < -3 was associated with mortality (P < 0.001) independent of CD4% and < -2 was associated with immunological failure (P ≤ 0.03) independent of age at cART.
CONCLUSIONS: Weight monitoring provides useful data to inform clinical management of children on cART in resource-limited settings.
Methods: Study end points were as follows: (1) a CD4 count <200 cells/mm3 followed by a CD4 count ≥200 cells/mm3 (transient CD4 <200); (2) CD4 count <200 cells/mm3 confirmed within 6 months (confirmed CD4 <200); and (3) a new or recurrent World Health Organization (WHO) stage 3 or 4 illness (clinical failure). Kaplan-Meier curves and Cox regression were used to evaluate rates and predictors of transient CD4 <200, confirmed CD4 <200, and clinical failure among virally suppressed children aged 5-15 years who were enrolled in the TREAT Asia Pediatric HIV Observational Database.
Results: Data from 967 children were included in the analysis. At the time of confirmed viral suppression, median age was 10.2 years, 50.4% of children were female, and 95.4% were perinatally infected with HIV. Median CD4 cell count was 837 cells/mm3, and 54.8% of children were classified as having WHO stage 3 or 4 disease. In total, 18 transient CD4 <200 events, 2 confirmed CD4 <200 events, and10 clinical failures occurred at rates of 0.73 (95% confidence interval [95% CI], 0.46-1.16), 0.08 (95% CI, 0.02-0.32), and 0.40 (95% CI, 0.22-0.75) events per 100 patient-years, respectively. CD4 <500 cells/mm3 at the time of viral suppression confirmation was associated with higher rates of both CD4 outcomes.
Conclusions: Regular CD4 testing may be unnecessary for virally suppressed children aged 5-15 years with CD4 ≥500 cells/mm3.
METHODS: HIV-infected adults enrolled in the TREAT Asia HIV Observational Database were eligible if they had an HIV RNA measurement documented at the time of ART initiation. The dataset was randomly split into a derivation data set (75% of patients) and a validation data set (25%). Factors associated with pre-treatment HIV RNA <100,000 copies/mL were evaluated by logistic regression adjusted for study site. A prediction model and prediction scores were created.
RESULTS: A total of 2592 patients were enrolled for the analysis. Median [interquartile range (IQR)] age was 35.8 (29.9-42.5) years; CD4 count was 147 (50-248) cells/mm3; and pre-treatment HIV RNA was 100,000 (34,045-301,075) copies/mL. Factors associated with pre-treatment HIV RNA <100,000 copies/mL were age <30 years [OR 1.40 vs. 41-50 years; 95% confidence interval (CI) 1.10-1.80, p = 0.01], body mass index >30 kg/m2(OR 2.4 vs. <18.5 kg/m2; 95% CI 1.1-5.1, p = 0.02), anemia (OR 1.70; 95% CI 1.40-2.10, p 350 cells/mm3(OR 3.9 vs. <100 cells/mm3; 95% CI 2.0-4.1, p 2000 cells/mm3(OR 1.7 vs. <1000 cells/mm3; 95% CI 1.3-2.3, p 25 yielded the sensitivity of 46.7%, specificity of 79.1%, positive predictive value of 67.7%, and negative predictive value of 61.2% for prediction of pre-treatment HIV RNA <100,000 copies/mL among derivation patients.
CONCLUSION: A model prediction for pre-treatment HIV RNA <100,000 copies/mL produced an area under the ROC curve of 0.70. A larger sample size for prediction model development as well as for model validation is warranted.
METHODS: Treatment modification was defined as a change of two antiretrovirals, a drug class change or treatment interruption (TI), all for >14 days. We assessed factors associated with CD4 changes and undetectable viral load (UVL <1,000 copies/ml) at 1 year after second-line failure using linear and logistic regression, respectively. Survival time was analysed using competing risk regression.
RESULTS: Of the 328 patients who failed second-line ART in our cohorts, 208 (63%) had a subsequent treatment modification. Compared with those who continued the failing regimen, the average CD4 cell increase was higher in patients who had a modification without TI (difference =77.5, 95% CI 35.3, 119.7) while no difference was observed among those with TI (difference =-5.3, 95% CI -67.3, 56.8). Compared with those who continued the failing regimen, the odds of achieving UVL was lower in patients with TI (OR=0.18, 95% CI 0.06, 0.60) and similar among those who had a modification without TI (OR=1.97, 95% CI 0.95, 4.10), with proportions of UVL 60%, 22% and 75%, respectively. Survival time was not affected by treatment modifications.
CONCLUSIONS: CD4 cell improvements were observed in those who had treatment modification without TI compared with those on the failing regimen. When no other options are available, maintaining the same failing ART combination provided better VL control than interrupting treatment.
METHODS: Adults living with HIV enrolled in a regional observational cohort in Asia who had initiated combination antiretroviral therapy (cART) were included in the analysis. Factors associated with new TB diagnoses after cohort entry and survival after cART initiation were analysed using Cox regression, stratified by site.
RESULTS: A total of 7355 patients from 12 countries enrolled into the cohort between 2003 and 2016 were included in the study. There were 368 reported cases of TB after cohort entry with an incidence rate of 0.99 per 100 person-years (/100 pys). Multivariate analyses adjusted for viral load (VL), CD4 count, body mass index (BMI) and cART duration showed that CTX reduced the hazard for new TB infection by 28% (HR 0.72, 95% CI l 0.56, 0.93). Mortality after cART initiation was 0.85/100 pys, with a median follow-up time of 4.63 years. Predictors of survival included age, female sex, hepatitis C co-infection, TB diagnosis, HIV VL, CD4 count and BMI.
CONCLUSIONS: CTX was associated with a reduction in the hazard for new TB infection but did not impact survival in our Asian cohort. The potential preventive effect of CTX against TB during periods of severe immunosuppression should be further explored.
METHODS: Patients initiating cART between 2006 and 2013 were included. TI was defined as stopping cART for >1 day. Treatment failure was defined as confirmed virological, immunological or clinical failure. Time to treatment failure during cART was analysed using Cox regression, not including periods off treatment. Covariables with P < 0.10 in univariable analyses were included in multivariable analyses, where P < 0.05 was considered statistically significant.
RESULTS: Of 4549 patients from 13 countries in Asia, 3176 (69.8%) were male and the median age was 34 years. A total of 111 (2.4%) had TIs due to AEs and 135 (3.0%) had TIs for other reasons. Median interruption times were 22 days for AE and 148 days for non-AE TIs. In multivariable analyses, interruptions >30 days were associated with failure (31-180 days HR = 2.66, 95%CI (1.70-4.16); 181-365 days HR = 6.22, 95%CI (3.26-11.86); and >365 days HR = 9.10, 95% CI (4.27-19.38), all P < 0.001, compared to 0-14 days). Reasons for previous TI were not statistically significant (P = 0.158).
CONCLUSIONS: Duration of interruptions of more than 30 days was the key factor associated with large increases in subsequent risk of treatment failure. If TI is unavoidable, its duration should be minimised to reduce the risk of failure after treatment resumption.
METHODS: The effect of geraniin on DENV-2 RNA synthesis in infected Vero cells was tested using quantitative RT-PCR. The in vivo efficacy of geraniin in inhibiting DENV-2 infection was then tested using BALB/c mice with geraniin administered at three different times. The differences in spleen to body weight ratio, DENV-2 RNA load and liver damage between the three treatment groups as compared to DENV-2 infected mice without geraniin administration were determined on day eight post-infection.
RESULTS: Quantitative RT-PCR confirmed the decrease in viral RNA synthesis of infected Vero cells when treated with geraniin. Geraniin seemed to provide a protective effect on infected BALB/c mice liver when given at 24 h pre- and 24 h post-infection as liver damage was observed to be very mild even though a significant reduction of DENV-2 RNA load in serum was not observed in these two treatment groups. However, when administered at 72 h post-infection, severe liver damage in the form of necrosis and haemorrhage had prevailed despite a substantial reduction of DENV-2 RNA load in serum.
CONCLUSIONS: Geraniin was found to be effective in reducing DENV-2 RNA load when administered at 72 h post-infection while earlier administration could prevent severe liver damage caused by DENV-2 infection. These results provide evidence that geraniin is a potential candidate for the development of anti-dengue drug.