mirror of
https://github.com/celogeek/go-comic-converter.git
synced 2025-05-24 15:52:38 +02:00
429 lines
8.2 KiB
Go
429 lines
8.2 KiB
Go
package epubimageprocessor
|
|
|
|
import (
|
|
"archive/zip"
|
|
"bytes"
|
|
"errors"
|
|
"fmt"
|
|
"image"
|
|
"image/color"
|
|
_ "image/jpeg"
|
|
_ "image/png"
|
|
"io"
|
|
"io/fs"
|
|
"os"
|
|
"path/filepath"
|
|
"sort"
|
|
"strings"
|
|
"sync"
|
|
|
|
"golang.org/x/image/font/gofont/gomonobold"
|
|
_ "golang.org/x/image/webp"
|
|
|
|
"github.com/celogeek/go-comic-converter/v2/internal/sortpath"
|
|
"github.com/fogleman/gg"
|
|
"github.com/golang/freetype/truetype"
|
|
"github.com/nwaples/rardecode/v2"
|
|
pdfimage "github.com/raff/pdfreader/image"
|
|
"github.com/raff/pdfreader/pdfread"
|
|
)
|
|
|
|
type task struct {
|
|
Id int
|
|
Image image.Image
|
|
Path string
|
|
Name string
|
|
Error error
|
|
}
|
|
|
|
var errNoImagesFound = errors.New("no images found")
|
|
|
|
// only accept jpg, png and webp as source file
|
|
func (e *EPUBImageProcessor) isSupportedImage(path string) bool {
|
|
switch strings.ToLower(filepath.Ext(path)) {
|
|
case ".jpg", ".jpeg", ".png", ".webp":
|
|
{
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// Load images from input
|
|
func (e *EPUBImageProcessor) load() (totalImages int, output chan *task, err error) {
|
|
fi, err := os.Stat(e.Input)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
// get all images though a channel of bytes
|
|
if fi.IsDir() {
|
|
return e.loadDir()
|
|
} else {
|
|
switch ext := strings.ToLower(filepath.Ext(e.Input)); ext {
|
|
case ".cbz", ".zip":
|
|
return e.loadCbz()
|
|
case ".cbr", ".rar":
|
|
return e.loadCbr()
|
|
case ".pdf":
|
|
return e.loadPdf()
|
|
default:
|
|
err = fmt.Errorf("unknown file format (%s): support .cbz, .zip, .cbr, .rar, .pdf", ext)
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
func (e *EPUBImageProcessor) corruptedImage(path, name string) image.Image {
|
|
var w, h float64 = 1200, 1920
|
|
f, _ := truetype.Parse(gomonobold.TTF)
|
|
face := truetype.NewFace(f, &truetype.Options{Size: 64, DPI: 72})
|
|
txt := name
|
|
if path != "" {
|
|
txt += "\nin " + filepath.Clean(path)
|
|
}
|
|
txt += "\nis corrupted!"
|
|
|
|
g := gg.NewContext(int(w), int(h))
|
|
g.SetColor(color.White)
|
|
g.Clear()
|
|
g.SetColor(color.Black)
|
|
g.DrawRoundedRectangle(0, 0, w, h, 0.5)
|
|
g.SetLineWidth(6)
|
|
g.Stroke()
|
|
g.DrawRoundedRectangle(0, 0, 480, 640, 0.5)
|
|
g.SetFontFace(face)
|
|
g.DrawStringWrapped(txt, w/2, h/2, 0.5, 0.5, 640, 1.5, gg.AlignCenter)
|
|
return g.Image()
|
|
}
|
|
|
|
// load a directory of images
|
|
func (e *EPUBImageProcessor) loadDir() (totalImages int, output chan *task, err error) {
|
|
images := make([]string, 0)
|
|
|
|
input := filepath.Clean(e.Input)
|
|
err = filepath.WalkDir(input, func(path string, d fs.DirEntry, err error) error {
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if !d.IsDir() && e.isSupportedImage(path) {
|
|
images = append(images, path)
|
|
}
|
|
return nil
|
|
})
|
|
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
totalImages = len(images)
|
|
|
|
if totalImages == 0 {
|
|
err = errNoImagesFound
|
|
return
|
|
}
|
|
|
|
sort.Sort(sortpath.By(images, e.SortPathMode))
|
|
|
|
// Queue all file with id
|
|
type job struct {
|
|
Id int
|
|
Path string
|
|
}
|
|
jobs := make(chan *job)
|
|
go func() {
|
|
defer close(jobs)
|
|
for i, path := range images {
|
|
jobs <- &job{i, path}
|
|
}
|
|
}()
|
|
|
|
// read in parallel and get an image
|
|
output = make(chan *task, e.Workers)
|
|
wg := &sync.WaitGroup{}
|
|
for j := 0; j < e.WorkersRatio(50); j++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
for job := range jobs {
|
|
var img image.Image
|
|
var err error
|
|
if !e.Dry {
|
|
var f *os.File
|
|
f, err = os.Open(job.Path)
|
|
if err == nil {
|
|
img, _, err = image.Decode(f)
|
|
f.Close()
|
|
}
|
|
}
|
|
|
|
p, fn := filepath.Split(job.Path)
|
|
if p == input {
|
|
p = ""
|
|
} else {
|
|
p = p[len(input)+1:]
|
|
}
|
|
if err != nil {
|
|
img = e.corruptedImage(p, fn)
|
|
}
|
|
output <- &task{
|
|
Id: job.Id,
|
|
Image: img,
|
|
Path: p,
|
|
Name: fn,
|
|
Error: err,
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
// wait all done and close
|
|
go func() {
|
|
wg.Wait()
|
|
close(output)
|
|
}()
|
|
|
|
return
|
|
}
|
|
|
|
// load a zip file that include images
|
|
func (e *EPUBImageProcessor) loadCbz() (totalImages int, output chan *task, err error) {
|
|
r, err := zip.OpenReader(e.Input)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
images := make([]*zip.File, 0)
|
|
for _, f := range r.File {
|
|
if !f.FileInfo().IsDir() && e.isSupportedImage(f.Name) {
|
|
images = append(images, f)
|
|
}
|
|
}
|
|
|
|
totalImages = len(images)
|
|
|
|
if totalImages == 0 {
|
|
r.Close()
|
|
err = errNoImagesFound
|
|
return
|
|
}
|
|
|
|
var names []string
|
|
for _, img := range images {
|
|
names = append(names, img.Name)
|
|
}
|
|
sort.Sort(sortpath.By(names, e.SortPathMode))
|
|
|
|
indexedNames := make(map[string]int)
|
|
for i, name := range names {
|
|
indexedNames[name] = i
|
|
}
|
|
|
|
type job struct {
|
|
Id int
|
|
F *zip.File
|
|
}
|
|
jobs := make(chan *job)
|
|
go func() {
|
|
defer close(jobs)
|
|
for _, img := range images {
|
|
jobs <- &job{indexedNames[img.Name], img}
|
|
}
|
|
}()
|
|
|
|
output = make(chan *task, e.Workers)
|
|
wg := &sync.WaitGroup{}
|
|
for j := 0; j < e.WorkersRatio(50); j++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
for job := range jobs {
|
|
var img image.Image
|
|
var err error
|
|
if !e.Dry {
|
|
var f io.ReadCloser
|
|
f, err = job.F.Open()
|
|
if err == nil {
|
|
img, _, err = image.Decode(f)
|
|
}
|
|
f.Close()
|
|
}
|
|
|
|
p, fn := filepath.Split(filepath.Clean(job.F.Name))
|
|
if err != nil {
|
|
img = e.corruptedImage(p, fn)
|
|
}
|
|
output <- &task{
|
|
Id: job.Id,
|
|
Image: img,
|
|
Path: p,
|
|
Name: fn,
|
|
Error: err,
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
go func() {
|
|
wg.Wait()
|
|
close(output)
|
|
r.Close()
|
|
}()
|
|
return
|
|
}
|
|
|
|
// load a rar file that include images
|
|
func (e *EPUBImageProcessor) loadCbr() (totalImages int, output chan *task, err error) {
|
|
var isSolid bool
|
|
files, err := rardecode.List(e.Input)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
names := make([]string, 0)
|
|
for _, f := range files {
|
|
if !f.IsDir && e.isSupportedImage(f.Name) {
|
|
if f.Solid {
|
|
isSolid = true
|
|
}
|
|
names = append(names, f.Name)
|
|
}
|
|
}
|
|
|
|
totalImages = len(names)
|
|
if totalImages == 0 {
|
|
err = errNoImagesFound
|
|
return
|
|
}
|
|
|
|
sort.Sort(sortpath.By(names, e.SortPathMode))
|
|
|
|
indexedNames := make(map[string]int)
|
|
for i, name := range names {
|
|
indexedNames[name] = i
|
|
}
|
|
|
|
type job struct {
|
|
Id int
|
|
Name string
|
|
Open func() (io.ReadCloser, error)
|
|
}
|
|
|
|
jobs := make(chan *job)
|
|
go func() {
|
|
defer close(jobs)
|
|
if isSolid && !e.Dry {
|
|
r, rerr := rardecode.OpenReader(e.Input)
|
|
if rerr != nil {
|
|
fmt.Fprintf(os.Stderr, "\nerror processing image %s: %s\n", e.Input, rerr)
|
|
os.Exit(1)
|
|
}
|
|
defer r.Close()
|
|
for {
|
|
f, rerr := r.Next()
|
|
if rerr != nil {
|
|
if rerr == io.EOF {
|
|
break
|
|
}
|
|
fmt.Fprintf(os.Stderr, "\nerror processing image %s: %s\n", f.Name, rerr)
|
|
os.Exit(1)
|
|
}
|
|
if i, ok := indexedNames[f.Name]; ok {
|
|
var b bytes.Buffer
|
|
_, rerr = io.Copy(&b, r)
|
|
if rerr != nil {
|
|
fmt.Fprintf(os.Stderr, "\nerror processing image %s: %s\n", f.Name, rerr)
|
|
os.Exit(1)
|
|
}
|
|
jobs <- &job{i, f.Name, func() (io.ReadCloser, error) {
|
|
return io.NopCloser(bytes.NewReader(b.Bytes())), nil
|
|
}}
|
|
}
|
|
}
|
|
} else {
|
|
for _, img := range files {
|
|
if i, ok := indexedNames[img.Name]; ok {
|
|
jobs <- &job{i, img.Name, img.Open}
|
|
}
|
|
}
|
|
}
|
|
}()
|
|
|
|
// send file to the queue
|
|
output = make(chan *task, e.Workers)
|
|
wg := &sync.WaitGroup{}
|
|
for j := 0; j < e.WorkersRatio(50); j++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
for job := range jobs {
|
|
var img image.Image
|
|
var err error
|
|
if !e.Dry {
|
|
var f io.ReadCloser
|
|
f, err = job.Open()
|
|
if err == nil {
|
|
img, _, err = image.Decode(f)
|
|
}
|
|
f.Close()
|
|
}
|
|
|
|
p, fn := filepath.Split(filepath.Clean(job.Name))
|
|
if err != nil {
|
|
img = e.corruptedImage(p, fn)
|
|
}
|
|
output <- &task{
|
|
Id: job.Id,
|
|
Image: img,
|
|
Path: p,
|
|
Name: fn,
|
|
Error: err,
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
go func() {
|
|
wg.Wait()
|
|
close(output)
|
|
}()
|
|
return
|
|
}
|
|
|
|
// extract image from a pdf
|
|
func (e *EPUBImageProcessor) loadPdf() (totalImages int, output chan *task, err error) {
|
|
pdf := pdfread.Load(e.Input)
|
|
if pdf == nil {
|
|
err = fmt.Errorf("can't read pdf")
|
|
return
|
|
}
|
|
|
|
totalImages = len(pdf.Pages())
|
|
pageFmt := fmt.Sprintf("page %%0%dd", len(fmt.Sprintf("%d", totalImages)))
|
|
output = make(chan *task)
|
|
go func() {
|
|
defer close(output)
|
|
defer pdf.Close()
|
|
for i := 0; i < totalImages; i++ {
|
|
var img image.Image
|
|
var err error
|
|
if !e.Dry {
|
|
img, err = pdfimage.Extract(pdf, i+1)
|
|
}
|
|
|
|
name := fmt.Sprintf(pageFmt, i+1)
|
|
if err != nil {
|
|
img = e.corruptedImage("", name)
|
|
}
|
|
output <- &task{
|
|
Id: i,
|
|
Image: img,
|
|
Path: "",
|
|
Name: name,
|
|
Error: err,
|
|
}
|
|
}
|
|
}()
|
|
|
|
return
|
|
}
|