I have been looking the site and found that the [ links ] labels runs some javascript at onclick event that sends you to an intermediate site, page etc. Thus so far it is not easy to scrap from them.
I found this solution that matches 65 of the 68 links written as text in the "#article-back" section. It seems three links are not well formatted thus not matched (i.e. "h ttp://"). I hope it has been helpful.
Edit:
Regexp taken from this answer
library(rvest)
library(dplyr)
link <- "https://www.scielo.org.mx/scielo.php?script=sci_arttext&pid=S2448-76782022000100004&lang=es"
page <- read_html(link)
text <- page %>% html_node("#article-back") %>%
html_text()
matches <- gregexpr(
"\\b(https?|ftp|file)://)?[-A-Za-z0-9+&@#/%?=~_|!:,.;]+[-A-Za-z0-9+&@#/%=~_|]",
links)
links <- regmatches(links, matches)
Edit 2
For scrap from the javascript in onclick:
library(rvest)
library(dplyr)
link <- "https://www.scielo.org.mx/scielo.php?script=sci_arttext&pid=S2448-76782022000100004&lang=es"
page <- read_html(link)
text <- page %>% html_node("#article-back") %>%
html_nodes("a") %>% html_attr("onclick")
links <- gsub(".*(/[^']+).*", "https://www.scielo.org.mx/\\1", text[!is.na(text)])
links_pid <- gsub(".*pid=([^&]+)&.*", "\\1", links)
links_pid