METHODS: We investigated serum creatinine (S-Cr) monitoring rates before and during ART and the incidence and prevalence of renal dysfunction after starting TDF by using data from a regional cohort of HIV-infected individuals in the Asia-Pacific. Time to renal dysfunction was defined as time from TDF initiation to the decline in estimated glomerular filtration rate (eGFR) to <60 ml/min/1.73m2 with >30% reduction from baseline using the Chronic Kidney Disease Epidemiology Collaboration (CKD-EPI) equation or the decision to stop TDF for reported TDF-nephrotoxicity. Predictors of S-Cr monitoring rates were assessed by Poisson regression and risk factors for developing renal dysfunction were assessed by Cox regression.
RESULTS: Among 2,425 patients who received TDF, S-Cr monitoring rates increased from 1.01 to 1.84 per person per year after starting TDF (incidence rate ratio 1.68, 95%CI 1.62-1.74, p <0.001). Renal dysfunction on TDF occurred in 103 patients over 5,368 person-years of TDF use (4.2%; incidence 1.75 per 100 person-years). Risk factors for developing renal dysfunction included older age (>50 vs. ≤30, hazard ratio [HR] 5.39, 95%CI 2.52-11.50, p <0.001; and using PI-based regimen (HR 1.93, 95%CI 1.22-3.07, p = 0.005). Having an eGFR prior to TDF (pre-TDF eGFR) of ≥60 ml/min/1.73m2 showed a protective effect (HR 0.38, 95%CI, 0.17-0.85, p = 0.018).
CONCLUSIONS: Renal dysfunction on commencing TDF use was not common, however, older age, lower baseline eGFR and PI-based ART were associated with higher risk of renal dysfunction during TDF use in adult HIV-infected individuals in the Asia-Pacific region.
AREAS COVERED: Using the publicly available IVAC VIEW-hub platform, we reviewed 52 studies on vaccine effectiveness (VE) after booster vaccinations. VE were reported for SARS-CoV-2 symptomatic infection, severe disease and death and stratified by vaccine schedule and age. In addition, a non-systematic literature review of safety was performed to identify single or multi-country studies investigating adverse event rates for at least two of the currently available COVID-19 vaccines.
EXPERT OPINION: Booster shots of the current COVID-19 vaccines provide consistently high protection against Omicron-related severe disease and death. Additionally, this protection appears to be conserved for at least 3 months, with a small but significant waning after that. The positive risk-benefit ratio of these vaccines is well established, giving us confidence to administer additional doses as required. Future vaccination strategies will likely include a combination of schedules based on risk profile, as overly frequent boosting may be neither beneficial nor sustainable for the general population.
METHODS: PLHIV enrolled in the Therapeutics, Research, Education and AIDS Training in Asia (TREAT Asia) HIV Observational Database (TAHOD) who initiated ART with a CD4 count 1 year were censored at 12 months. Competing risk regression was used to analyse risk factors with loss to follow-up as a competing risk.
RESULTS: A total of 1813 PLHIV were included in the study, of whom 74% were male. With 73 (4%) deaths, the overall first-year mortality rate was 4.27 per 100 person-years (PY). Thirty-eight deaths (52%) were AIDS-related, 10 (14%) were immune reconstituted inflammatory syndrome (IRIS)-related, 13 (18%) were non-AIDS-related and 12 (16%) had an unknown cause. Risk factors included having a body mass index (BMI) 100 cells/μL: SHR 0.12; 95% CI 0.05-0.26) was associated with reduced hazard for mortality compared to CD4 count ≤ 25 cells/μL.
CONCLUSIONS: Fifty-two per cent of early deaths were AIDS-related. Efforts to initiate ART at CD4 counts > 50 cell/μL are associated with improved short-term survival rates, even in those with late stages of HIV disease.
METHODS AND FINDINGS: We reviewed all GenBank submissions of HIV-1 reverse transcriptase sequences with or without protease and identified 287 studies published between March 1, 2000, and December 31, 2013, with more than 25 recently or chronically infected ARV-naïve individuals. These studies comprised 50,870 individuals from 111 countries. Each set of study sequences was analyzed for phylogenetic clustering and the presence of 93 surveillance drug-resistance mutations (SDRMs). The median overall TDR prevalence in sub-Saharan Africa (SSA), south/southeast Asia (SSEA), upper-income Asian countries, Latin America/Caribbean, Europe, and North America was 2.8%, 2.9%, 5.6%, 7.6%, 9.4%, and 11.5%, respectively. In SSA, there was a yearly 1.09-fold (95% CI: 1.05-1.14) increase in odds of TDR since national ARV scale-up attributable to an increase in non-nucleoside reverse transcriptase inhibitor (NNRTI) resistance. The odds of NNRTI-associated TDR also increased in Latin America/Caribbean (odds ratio [OR] = 1.16; 95% CI: 1.06-1.25), North America (OR = 1.19; 95% CI: 1.12-1.26), Europe (OR = 1.07; 95% CI: 1.01-1.13), and upper-income Asian countries (OR = 1.33; 95% CI: 1.12-1.55). In SSEA, there was no significant change in the odds of TDR since national ARV scale-up (OR = 0.97; 95% CI: 0.92-1.02). An analysis limited to sequences with mixtures at less than 0.5% of their nucleotide positions—a proxy for recent infection—yielded trends comparable to those obtained using the complete dataset. Four NNRTI SDRMs—K101E, K103N, Y181C, and G190A—accounted for >80% of NNRTI-associated TDR in all regions and subtypes. Sixteen nucleoside reverse transcriptase inhibitor (NRTI) SDRMs accounted for >69% of NRTI-associated TDR in all regions and subtypes. In SSA and SSEA, 89% of NNRTI SDRMs were associated with high-level resistance to nevirapine or efavirenz, whereas only 27% of NRTI SDRMs were associated with high-level resistance to zidovudine, lamivudine, tenofovir, or abacavir. Of 763 viruses with TDR in SSA and SSEA, 725 (95%) were genetically dissimilar; 38 (5%) formed 19 sequence pairs. Inherent limitations of this study are that some cohorts may not represent the broader regional population and that studies were heterogeneous with respect to duration of infection prior to sampling.
CONCLUSIONS: Most TDR strains in SSA and SSEA arose independently, suggesting that ARV regimens with a high genetic barrier to resistance combined with improved patient adherence may mitigate TDR increases by reducing the generation of new ARV-resistant strains. A small number of NNRTI-resistance mutations were responsible for most cases of high-level resistance, suggesting that inexpensive point-mutation assays to detect these mutations may be useful for pre-therapy screening in regions with high levels of TDR. In the context of a public health approach to ARV therapy, a reliable point-of-care genotypic resistance test could identify which patients should receive standard first-line therapy and which should receive a protease-inhibitor-containing regimen.
METHODS: HIV+ patients from the Australian HIV Observational Database (AHOD) and the TREAT Asia HIV Observational Database (TAHOD) meeting specific criteria were included. In these analyses Asian and Caucasian status were defined by cohort. Factors associated with a low CD4:CD8 ratio (cutoff <0.2) prior to ART commencement, and with achieving a normal CD4:CD8 ratio (>1) at 12 and 24 months post ART commencement were assessed using logistic regression.
RESULTS: There were 591 patients from AHOD and 2,620 patients from TAHOD who met the inclusion criteria. TAHOD patients had a significantly (P<0.001) lower odds of having a baseline (prior to ART initiation) CD4:CD8 ratio greater than 0.2. After 12 months of ART, AHOD patients were more than twice as likely to achieve a normal CD4:CD8 ratio compared to TAHOD patients (15% versus 6%). However, after adjustment for confounding factors there was no significant difference between cohorts in the odds of achieving a CD4:CD8 ratio >1 (P=0.475).
CONCLUSIONS: We found a significantly lower CD4:CD8 ratio prior to commencing ART in TAHOD compared to AHOD even after adjusting for confounders. However, after adjustment, there was no significant difference between the cohorts in odds of achieving normal ratio. Baseline CD4+ and CD8+ counts seem to be the main driver for this difference between these two populations.
METHODS: Patients from TREAT Asia HIV Observational Database (September 2015 data transfer) aged 18 years and older with a CD4 count <50 cells/mm at ART initiation were included. The effect of macrolide prophylaxis on HIV-associated mortality or AIDS-defining conditions (as a combined outcome) and HIV-associated mortality alone were evaluated using competing risk regression. Sensitivity analysis was conducted in patients with a CD4 <100 cells/mm at ART initiation.
RESULTS: Of 1345 eligible patients, 10.6% received macrolide prophylaxis. The rate of the combined outcome was 7.35 [95% confidence interval (CI): 6.04 to 8.95] per 100 patient-years, whereas the rate of HIV-associated mortality was 3.14 (95% CI: 2.35 to 4.19) per 100 patient-years. Macrolide use was associated with a significantly decreased risk of HIV-associated mortality (hazard ratio 0.10, 95% CI: 0.01 to 0.80, P = 0.031) but not with the combined outcome (hazard ratio 0.86, 95% CI: 0.32 to 2.229, P = 0.764). Sensitivity analyses showed consistent results among patients with a CD4 <100 cells/mm at ART initiation.
CONCLUSIONS: Macrolide prophylaxis is associated with improved survival among Asian HIV-infected patients with low CD4 cell counts and on ART. This study suggests the increased usage and coverage of macrolide prophylaxis among people living with HIV in Asia.
METHODS: Adults living with HIV enrolled in a regional observational cohort in Asia who had initiated combination antiretroviral therapy (cART) were included in the analysis. Factors associated with new TB diagnoses after cohort entry and survival after cART initiation were analysed using Cox regression, stratified by site.
RESULTS: A total of 7355 patients from 12 countries enrolled into the cohort between 2003 and 2016 were included in the study. There were 368 reported cases of TB after cohort entry with an incidence rate of 0.99 per 100 person-years (/100 pys). Multivariate analyses adjusted for viral load (VL), CD4 count, body mass index (BMI) and cART duration showed that CTX reduced the hazard for new TB infection by 28% (HR 0.72, 95% CI l 0.56, 0.93). Mortality after cART initiation was 0.85/100 pys, with a median follow-up time of 4.63 years. Predictors of survival included age, female sex, hepatitis C co-infection, TB diagnosis, HIV VL, CD4 count and BMI.
CONCLUSIONS: CTX was associated with a reduction in the hazard for new TB infection but did not impact survival in our Asian cohort. The potential preventive effect of CTX against TB during periods of severe immunosuppression should be further explored.
METHODS: HIV-infected adults enrolled in the TREAT Asia HIV Observational Database were eligible if they had an HIV RNA measurement documented at the time of ART initiation. The dataset was randomly split into a derivation data set (75% of patients) and a validation data set (25%). Factors associated with pre-treatment HIV RNA <100,000 copies/mL were evaluated by logistic regression adjusted for study site. A prediction model and prediction scores were created.
RESULTS: A total of 2592 patients were enrolled for the analysis. Median [interquartile range (IQR)] age was 35.8 (29.9-42.5) years; CD4 count was 147 (50-248) cells/mm3; and pre-treatment HIV RNA was 100,000 (34,045-301,075) copies/mL. Factors associated with pre-treatment HIV RNA <100,000 copies/mL were age <30 years [OR 1.40 vs. 41-50 years; 95% confidence interval (CI) 1.10-1.80, p = 0.01], body mass index >30 kg/m2(OR 2.4 vs. <18.5 kg/m2; 95% CI 1.1-5.1, p = 0.02), anemia (OR 1.70; 95% CI 1.40-2.10, p 350 cells/mm3(OR 3.9 vs. <100 cells/mm3; 95% CI 2.0-4.1, p 2000 cells/mm3(OR 1.7 vs. <1000 cells/mm3; 95% CI 1.3-2.3, p 25 yielded the sensitivity of 46.7%, specificity of 79.1%, positive predictive value of 67.7%, and negative predictive value of 61.2% for prediction of pre-treatment HIV RNA <100,000 copies/mL among derivation patients.
CONCLUSION: A model prediction for pre-treatment HIV RNA <100,000 copies/mL produced an area under the ROC curve of 0.70. A larger sample size for prediction model development as well as for model validation is warranted.
METHODS: Data from two regional cohort observational databases were analyzed for trends in median CD4 cell counts at ART initiation and the proportion of late ART initiation (CD4 cell counts <200 cells/mm(3) or prior AIDS diagnosis). Predictors for late ART initiation and mortality were determined.
RESULTS: A total of 2737 HIV-positive ART-naïve patients from 22 sites in 13 Asian countries and territories were eligible. The overall median (IQR) CD4 cell count at ART initiation was 150 (46-241) cells/mm(3). Median CD4 cell counts at ART initiation increased over time, from a low point of 115 cells/mm(3) in 2008 to a peak of 302 cells/mm(3) after 2011 (p for trend 0.002). The proportion of patients with late ART initiation significantly decreased over time from 79.1% before 2007 to 36.3% after 2011 (p for trend <0.001). Factors associated with late ART initiation were year of ART initiation (e.g. 2010 vs. before 2007; OR 0.40, 95% CI 0.27-0.59; p<0.001), sex (male vs. female; OR 1.51, 95% CI 1.18-1.93; p=0.001) and HIV exposure risk (heterosexual vs. homosexual; OR 1.66, 95% CI 1.24-2.23; p=0.001 and intravenous drug use vs. homosexual; OR 3.03, 95% CI 1.77-5.21; p<0.001). Factors associated with mortality after ART initiation were late ART initiation (HR 2.13, 95% CI 1.19-3.79; p=0.010), sex (male vs. female; HR 2.12, 95% CI 1.31-3.43; p=0.002), age (≥51 vs. ≤30 years; HR 3.91, 95% CI 2.18-7.04; p<0.001) and hepatitis C serostatus (positive vs. negative; HR 2.48, 95% CI 1.-4.36; p=0.035).
CONCLUSIONS: Median CD4 cell count at ART initiation among Asian patients significantly increases over time but the proportion of patients with late ART initiation is still significant. ART initiation at higher CD4 cell counts remains a challenge. Strategic interventions to increase earlier diagnosis of HIV infection and prompt more rapid linkage to ART must be implemented.
METHODS: This study included people living with HIV enrolled in a longitudinal cohort study from 2003 to 2019, receiving antiretroviral therapy (ART), and without prior tuberculosis. BMI at ART initiation was categorized using Asian BMI classifications: underweight (<18.5 kg/m2 ), normal (18.5-22.9 kg/m2 ), overweight (23-24.9 kg/m2 ), and obese (≥25 kg/m2 ). High FBG was defined as a single post-ART FBG measurement ≥126 mg/dL. Factors associated with high FBG were analyzed using Cox regression models stratified by site.
RESULTS: A total of 3939 people living with HIV (63% male) were included. In total, 50% had a BMI in the normal weight range, 23% were underweight, 13% were overweight, and 14% were obese. Median age at ART initiation was 34 years (interquartile range 29-41). Overall, 8% had a high FBG, with an incidence rate of 1.14 per 100 person-years. Factors associated with an increased hazard of high FBG included being obese (≥25 kg/m2 ) compared with normal weight (hazard ratio [HR] = 1.79; 95% confidence interval [CI] 1.31-2.44; p 25 kg/m2 were at increased risk of high FBG. This indicates that regular assessments should be performed in those with high BMI, irrespective of the classification used.
METHODS: The study population consisted of HIV-infected patients enrolled in the TREAT Asia HIV Observational Database (TAHOD). Individuals were included in this analysis if they started combination antiretroviral treatment (cART) after 2002, were being treated at a centre that documented a median rate of viral load monitoring ≥0.8 tests/patient/year among TAHOD enrolees, and experienced a minor or major treatment substitution while on virally suppressive cART. The primary endpoint to evaluate outcomes was clinical or virological failure (VF), followed by an ART class change. Clinical failure was defined as death or an AIDS diagnosis. VF was defined as confirmed viral load measurements ≥400 copies/mL followed by an ART class change within six months. Minor regimen substitutions were defined as within-class changes and major regimen substitutions were defined as changes to a drug class. The patterns of substitutions and rate of clinical or VF after substitutions were analyzed.
RESULTS: Of 3994 adults who started ART after 2002, 3119 (78.1%) had at least one period of virological suppression. Among these, 1170 (37.5%) underwent a minor regimen substitution, and 296 (9.5%) underwent a major regimen substitution during suppression. The rates of clinical or VF were 1.48/100 person years (95% CI 1.14 to 1.91) in the minor substitution group, 2.85/100 person years (95% CI 1.88 to 4.33) in the major substitution group and 2.53/100 person years (95% CI 2.20 to 2.92) among patients that did not undergo a treatment substitution.
CONCLUSIONS: The rate of clinical or VF was low in both major and minor substitution groups, showing that regimen substitution is generally effective in non-clinical trial settings in Asian countries.
METHODS: We used data from the TREAT Asia HIV Observational Database. Patients were included if they started antiretroviral therapy during or after 2003, had a serum creatinine measurement at antiretroviral therapy initiation (baseline), and had at least 2 follow-up creatinine measurements taken ≥3 months apart. Patients with a baseline estimated glomerular filtration rate (eGFR) ≤60 mL/min/1.73 m2 were excluded. Chronic kidney disease was defined as 2 consecutive eGFR values ≤60 mL/min/1.73 m2 taken ≥3 months apart. Generalized estimating equations were used to identify factors associated with eGFR change. Competing risk regression adjusted for study site, age and sex, and cumulative incidence plots were used to evaluate factors associated with chronic kidney disease (CKD).
RESULTS: Of 2547 patients eligible for this analysis, tenofovir was being used by 703 (27.6%) at baseline. Tenofovir use, high baseline eGFR, advanced HIV disease stage, and low nadir CD4 were associated with a decrease in eGFR during follow-up. Chronic kidney disease occurred at a rate of 3.4 per 1000 patient/years. Factors associated with CKD were tenofovir use, old age, low baseline eGFR, low nadir CD4, and protease inhibitor use.
CONCLUSIONS: There is an urgent need to enhance renal monitoring and management capacity among at-risk groups in Asia and improve access to less nephrotoxic antiretrovirals.
METHODS: Incidence of malignancy after cohort enrollment was evaluated. Factors associated with development of hematological and nonhematological malignancy were analyzed using competing risk regression and survival time using Kaplan-Meier.
RESULTS: Of 7455 patients, 107 patients (1%) developed a malignancy: 34 (0.5%) hematological [0.08 per 100 person-years (/100PY)] and 73 (1%) nonhematological (0.17/100PY). Of the hematological malignancies, non-Hodgkin lymphoma was predominant (n = 26, 76%): immunoblastic (n = 6, 18%), Burkitt (n = 5, 15%), diffuse large B-cell (n = 5, 15%), and unspecified (n = 10, 30%). Others include central nervous system lymphoma (n = 7, 21%) and myelodysplastic syndrome (n = 1, 3%). Nonhematological malignancies were mostly Kaposi sarcoma (n = 12, 16%) and cervical cancer (n = 10, 14%). Risk factors for hematological malignancy included age >50 vs. ≤30 years [subhazard ratio (SHR) = 6.48, 95% confidence interval (CI): 1.79 to 23.43] and being from a high-income vs. a lower-middle-income country (SHR = 3.97, 95% CI: 1.45 to 10.84). Risk was reduced with CD4 351-500 cells/µL (SHR = 0.20, 95% CI: 0.05 to 0.74) and CD4 >500 cells/µL (SHR = 0.14, 95% CI: 0.04 to 0.78), compared to CD4 ≤200 cells/µL. Similar risk factors were seen for nonhematological malignancy, with prior AIDS diagnosis showing a weak association. Patients diagnosed with a hematological malignancy had shorter survival time compared to patients diagnosed with a nonhematological malignancy.
CONCLUSIONS: Nonhematological malignancies were common but non-Hodgkin lymphoma was more predominant in our cohort. PLHIV from high-income countries were more likely to be diagnosed, indicating a potential underdiagnosis of cancer in low-income settings.
METHODS: Long-term LTFU was defined as LTFU occurring after 5 years on ART. LTFU was defined as (1) patients not seen in the previous 12 months; and (2) patients not seen in the previous 6 months. Factors associated with LTFU were analysed using competing risk regression.
RESULTS: Under the 12-month definition, the LTFU rate was 2.0 per 100 person-years (PY) [95% confidence interval (CI) 1.8-2.2 among 4889 patients included in the study. LTFU was associated with age > 50 years [sub-hazard ratio (SHR) 1.64; 95% CI 1.17-2.31] compared with 31-40 years, viral load ≥ 1000 copies/mL (SHR 1.86; 95% CI 1.16-2.97) compared with viral load < 1000 copies/mL, and hepatitis C coinfection (SHR 1.48; 95% CI 1.06-2.05). LTFU was less likely to occur in females, in individuals with higher CD4 counts, in those with self-reported adherence ≥ 95%, and in those living in high-income countries. The 6-month LTFU definition produced an incidence rate of 3.2 per 100 PY (95% CI 2.9-3.4 and had similar associations but with greater risks of LTFU for ART initiation in later years (2006-2009: SHR 2.38; 95% CI 1.93-2.94; and 2010-2011: SHR 4.26; 95% CI 3.17-5.73) compared with 2003-2005.
CONCLUSIONS: The long-term LTFU rate in our cohort was low, with older age being associated with LTFU. The increased risk of LTFU with later years of ART initiation in the 6-month analysis, but not the 12-month analysis, implies that there was a possible move towards longer HIV clinic scheduling in Asia.
METHODS: Treatment modification was defined as a change of two antiretrovirals, a drug class change or treatment interruption (TI), all for >14 days. We assessed factors associated with CD4 changes and undetectable viral load (UVL <1,000 copies/ml) at 1 year after second-line failure using linear and logistic regression, respectively. Survival time was analysed using competing risk regression.
RESULTS: Of the 328 patients who failed second-line ART in our cohorts, 208 (63%) had a subsequent treatment modification. Compared with those who continued the failing regimen, the average CD4 cell increase was higher in patients who had a modification without TI (difference =77.5, 95% CI 35.3, 119.7) while no difference was observed among those with TI (difference =-5.3, 95% CI -67.3, 56.8). Compared with those who continued the failing regimen, the odds of achieving UVL was lower in patients with TI (OR=0.18, 95% CI 0.06, 0.60) and similar among those who had a modification without TI (OR=1.97, 95% CI 0.95, 4.10), with proportions of UVL 60%, 22% and 75%, respectively. Survival time was not affected by treatment modifications.
CONCLUSIONS: CD4 cell improvements were observed in those who had treatment modification without TI compared with those on the failing regimen. When no other options are available, maintaining the same failing ART combination provided better VL control than interrupting treatment.