move fetchPage to own function, defer resp.Body.Close

This commit is contained in:
2026-02-08 21:40:25 -08:00
parent ad455a36e8
commit 349c28e29c

View File

@@ -12,6 +12,8 @@ import (
"time" "time"
"muzi/db" "muzi/db"
"github.com/jackc/pgx/v5"
) )
type LastFMTrack struct { type LastFMTrack struct {
@@ -60,6 +62,40 @@ type Response struct {
} `json:"recenttracks"` } `json:"recenttracks"`
} }
func fetchPage(client *http.Client, page int, lfmUsername, apiKey string, userId int) pageResult {
resp, err := client.Get(
"https://ws.audioscrobbler.com/2.0/?method=user.getrecenttracks&user=" +
lfmUsername + "&api_key=" + apiKey + "&format=json&limit=100&page=" + strconv.Itoa(page),
)
if err != nil {
return pageResult{pageNum: page, err: err}
}
defer resp.Body.Close()
var data Response
if err := json.NewDecoder(resp.Body).Decode(&data); err != nil {
return pageResult{pageNum: page, err: err}
}
var pageTracks []LastFMTrack
for j := range data.Recenttracks.Track {
if data.Recenttracks.Track[j].Attr.Nowplaying == "true" {
continue
}
unixTime, err := strconv.ParseInt(data.Recenttracks.Track[j].Date.Uts, 10, 64)
if err != nil {
continue
}
pageTracks = append(pageTracks, LastFMTrack{
UserId: userId,
Timestamp: time.Unix(unixTime, 0),
SongName: data.Recenttracks.Track[j].Name,
Artist: data.Recenttracks.Track[j].Artist.Text,
Album: data.Recenttracks.Track[j].Album.Text,
})
}
return pageResult{pageNum: page, tracks: pageTracks, err: nil}
}
func ImportLastFM( func ImportLastFM(
username string, username string,
apiKey string, apiKey string,
@@ -83,10 +119,15 @@ func ImportLastFM(
} }
return err return err
} }
defer resp.Body.Close()
var initialData Response var initialData Response
json.NewDecoder(resp.Body).Decode(&initialData) err = json.NewDecoder(resp.Body).Decode(&initialData)
if err != nil {
fmt.Fprintf(os.Stderr,
"Error decoding initial LastFM response: %v\n", err)
return err
}
totalPages, err := strconv.Atoi(initialData.Recenttracks.Attr.TotalPages) totalPages, err := strconv.Atoi(initialData.Recenttracks.Attr.TotalPages)
resp.Body.Close()
if err != nil { if err != nil {
fmt.Fprintf(os.Stderr, "Error parsing total pages: %v\n", err) fmt.Fprintf(os.Stderr, "Error parsing total pages: %v\n", err)
if progressChan != nil { if progressChan != nil {
@@ -115,40 +156,7 @@ func ImportLastFM(
go func(workerID int) { go func(workerID int) {
defer wg.Done() defer wg.Done()
for page := workerID + 1; page <= totalPages; page += 10 { for page := workerID + 1; page <= totalPages; page += 10 {
resp, err := client.Get( pageChan <- fetchPage(client, page, lfmUsername, apiKey, userId)
"https://ws.audioscrobbler.com/2.0/?method=user.getrecenttracks&user=" +
username + "&api_key=" + apiKey + "&format=json&limit=100&page=" + strconv.Itoa(page),
)
if err != nil {
pageChan <- pageResult{pageNum: page, err: err}
continue
}
var data Response
if err := json.NewDecoder(resp.Body).Decode(&data); err != nil {
resp.Body.Close()
pageChan <- pageResult{pageNum: page, err: err}
continue
}
resp.Body.Close()
var pageTracks []LastFMTrack
for j := range data.Recenttracks.Track {
if data.Recenttracks.Track[j].Attr.Nowplaying == "true" {
continue
}
unixTime, err := strconv.ParseInt(data.Recenttracks.Track[j].Date.Uts, 10, 64)
if err != nil {
continue
}
pageTracks = append(pageTracks, LastFMTrack{
UserId: userId,
Timestamp: time.Unix(unixTime, 0),
SongName: data.Recenttracks.Track[j].Name,
Artist: data.Recenttracks.Track[j].Artist.Text,
Album: data.Recenttracks.Track[j].Album.Text,
})
}
pageChan <- pageResult{pageNum: page, tracks: pageTracks, err: nil}
} }
}(worker) }(worker)
} }
@@ -171,13 +179,14 @@ func ImportLastFM(
for len(trackBatch) >= batchSize { for len(trackBatch) >= batchSize {
batch := trackBatch[:batchSize] batch := trackBatch[:batchSize]
trackBatch = trackBatch[batchSize:] trackBatch = trackBatch[batchSize:]
err := insertBatch(batch, &totalImported, batchSize) err := insertBatch(batch, &totalImported)
if err != nil { if err != nil {
fmt.Fprintf(os.Stderr, "Batch insert failed: %v\n", err) // prevent logs being filled by duplicate warnings
if !strings.Contains(err.Error(), "duplicate") {
fmt.Fprintf(os.Stderr, "Batch insert failed: %v\n", err)
}
} }
} }
fmt.Printf("Processed page %d/%d\n", result.pageNum, totalPages)
// increment completed pages counter // increment completed pages counter
completedMu.Lock() completedMu.Lock()
completedPages++ completedPages++
@@ -197,9 +206,12 @@ func ImportLastFM(
} }
if len(trackBatch) > 0 { if len(trackBatch) > 0 {
err := insertBatch(trackBatch, &totalImported, batchSize) err := insertBatch(trackBatch, &totalImported)
if err != nil { if err != nil {
fmt.Fprintf(os.Stderr, "Final batch insert failed: %v\n", err) // prevent logs being filled by duplicate warnings
if !strings.Contains(err.Error(), "duplicate") {
fmt.Fprintf(os.Stderr, "Final batch insert failed: %v\n", err)
}
} }
} }
@@ -218,65 +230,21 @@ func ImportLastFM(
return nil return nil
} }
func insertBatch(tracks []LastFMTrack, totalImported *int, batchSize int) error { func insertBatch(tracks []LastFMTrack, totalImported *int) error {
tx, err := db.Pool.Begin(context.Background()) copyCount, err := db.Pool.CopyFrom(context.Background(),
if err != nil { pgx.Identifier{"history"},
return err []string{
} "user_id", "timestamp", "song_name", "artist", "album_name",
"ms_played", "platform",
var batchValues []string },
var batchArgs []any pgx.CopyFromSlice(len(tracks), func(i int) ([]any, error) {
t := tracks[i]
for i, track := range tracks { return []any{
batchValues = append(batchValues, fmt.Sprintf( t.UserId, t.Timestamp, t.SongName, t.Artist,
"($%d, $%d, $%d, $%d, $%d, $%d, $%d)", t.Album, 0, "lastfm",
len(batchArgs)+1, }, nil
len(batchArgs)+2, }),
len(batchArgs)+3, )
len(batchArgs)+4, *totalImported += int(copyCount)
len(batchArgs)+5, return err
len(batchArgs)+6,
len(batchArgs)+7,
))
// lastfm doesn't store playtime for each track, so set to 0
batchArgs = append(
batchArgs,
track.UserId,
track.Timestamp,
track.SongName,
track.Artist,
track.Album,
0,
"lastfm",
)
if len(batchValues) >= batchSize || i == len(tracks)-1 {
result, err := tx.Exec(
context.Background(),
`INSERT INTO history (user_id, timestamp, song_name, artist, album_name, ms_played, platform) VALUES `+
strings.Join(
batchValues,
", ",
)+` ON CONFLICT ON CONSTRAINT history_user_id_song_name_artist_timestamp_key DO NOTHING;`,
batchArgs...,
)
if err != nil {
tx.Rollback(context.Background())
return err
}
rowsAffected := result.RowsAffected()
if rowsAffected > 0 {
*totalImported += int(rowsAffected)
}
batchValues = batchValues[:0]
batchArgs = batchArgs[:0]
}
}
if err := tx.Commit(context.Background()); err != nil {
tx.Rollback(context.Background())
return err
}
return nil
} }