Remove noisy var _ = log.Printf lines.

Change-Id: Ia58b8ef5f271f542ae4fe61c7fb1497322770322
This commit is contained in:
Brad Fitzpatrick 2013-06-14 12:55:55 -07:00
parent a71162ab5f
commit ca58d8e2e0
22 changed files with 0 additions and 59 deletions

View File

@ -23,7 +23,6 @@ import (
"fmt"
"io"
"io/ioutil"
"log"
"os"
"path/filepath"
"strings"
@ -32,8 +31,6 @@ import (
"camlistore.org/pkg/osutil"
)
var _ = log.Printf
// TODO: rename StreamingFetcher to be Fetcher (the common case)
// TODO: add FetcherAt / FetchAt (for HTTP range requests). But then how

View File

@ -30,8 +30,6 @@ import (
"camlistore.org/pkg/jsonconfig"
)
var _ = log.Printf
const buffered = 8
type replicaStorage struct {

View File

@ -24,8 +24,6 @@ import (
"camlistore.org/pkg/blobserver"
)
var _ = log.Printf
var _ blobserver.MaxEnumerateConfig = (*s3Storage)(nil)
func (sto *s3Storage) MaxEnumerate() int { return 1000 }

View File

@ -18,13 +18,10 @@ package s3
import (
"io"
"log"
"camlistore.org/pkg/blobref"
)
var _ = log.Printf
func (sto *s3Storage) FetchStreaming(blob *blobref.BlobRef) (file io.ReadCloser, size int64, reterr error) {
file, size, reterr = sto.s3Client.Get(sto.bucket, blob.String())
return

View File

@ -22,16 +22,12 @@ import (
"hash"
"io"
"io/ioutil"
"log"
"os"
"camlistore.org/pkg/blobref"
"camlistore.org/pkg/blobserver"
//"camli/misc/amazon/s3"
)
var _ = log.Printf
const maxInMemorySlurp = 4 << 20 // 4MB. *shrug*
// amazonSlurper slurps up a blob to memory (or spilling to disk if

View File

@ -17,13 +17,9 @@ limitations under the License.
package s3
import (
"log"
"camlistore.org/pkg/blobref"
)
var _ = log.Printf
func (sto *s3Storage) RemoveBlobs(blobs []*blobref.BlobRef) error {
// TODO: do these in parallel
var reterr error

View File

@ -23,8 +23,6 @@ import (
"camlistore.org/pkg/blobref"
)
var _ = log.Printf
func (sto *s3Storage) StatBlobs(dest chan<- blobref.SizedBlobRef, blobs []*blobref.BlobRef, wait time.Duration) error {
// TODO: do n stats in parallel
for _, br := range blobs {

View File

@ -21,7 +21,6 @@ import (
"errors"
"fmt"
"io"
"log"
"regexp"
"camlistore.org/pkg/blobref"
@ -29,8 +28,6 @@ import (
"camlistore.org/pkg/schema"
)
var _ = log.Printf
func (c *Client) FetchSchemaBlob(b *blobref.BlobRef) (*schema.Blob, error) {
rc, _, err := c.FetchStreaming(b)
if err != nil {

View File

@ -34,8 +34,6 @@ import (
"camlistore.org/pkg/blobref"
)
var _ = log.Printf
var debugUploads = os.Getenv("CAMLI_DEBUG_UPLOADS") != ""
// multipartOverhead is how many extra bytes mime/multipart's

View File

@ -32,8 +32,6 @@ import (
"camlistore.org/pkg/types"
)
var _ = log.Printf
var ErrNotFound = errors.New("index: key not found")
type Storage interface {

View File

@ -35,8 +35,6 @@ import (
"camlistore.org/third_party/code.google.com/p/go.crypto/openpgp"
)
var _ = log.Printf
type EntityFetcher interface {
FetchEntity(keyId string) (*openpgp.Entity, error)
}

View File

@ -36,8 +36,6 @@ import (
"camlistore.org/third_party/code.google.com/p/go.crypto/openpgp"
)
var _ = log.Printf
const kMaxJSONLength = 1024 * 1024
type Handler struct {

View File

@ -22,7 +22,6 @@ import (
"encoding/json"
"errors"
"fmt"
"log"
"strings"
"camlistore.org/pkg/blobref"
@ -30,8 +29,6 @@ import (
"camlistore.org/third_party/code.google.com/p/go.crypto/openpgp/packet"
)
var _ = log.Printf
const sigSeparator = `,"camliSig":"`
// reArmor takes a camliSig (single line armor) and turns it back into an PGP-style

View File

@ -23,15 +23,12 @@ import (
"encoding/base64"
"fmt"
"io"
"log"
"net/http"
"sort"
"strings"
"time"
)
var _ = log.Printf
// See http://docs.amazonwebservices.com/AmazonS3/latest/dev/index.html?RESTAuthentication.html
type Auth struct {

View File

@ -25,15 +25,12 @@ import (
"hash"
"io"
"io/ioutil"
"log"
"net/http"
"net/url"
"os"
"strconv"
)
var _ = log.Printf
type Client struct {
*Auth
HttpClient *http.Client // or nil for default client

View File

@ -20,14 +20,11 @@ import (
"bufio"
"errors"
"fmt"
"log"
"os"
"os/exec"
"strings"
)
var _ = log.Printf
var ErrCancel = errors.New("pinentry: Cancel")
type Request struct {

View File

@ -19,7 +19,6 @@ package netutil
import (
"fmt"
"io/ioutil"
"log"
"net"
"net/http"
"net/http/httptest"
@ -29,8 +28,6 @@ import (
"time"
)
var _ = log.Printf
func TestLocalIPv4(t *testing.T) {
// Start listening on localhost IPv4, on some port.
ln, err := net.Listen("tcp", "127.0.0.1:0")

View File

@ -20,7 +20,6 @@ import (
"bytes"
"fmt"
"io/ioutil"
"log"
"math/rand"
"os"
"testing"
@ -28,8 +27,6 @@ import (
"camlistore.org/pkg/test"
)
var _ = log.Printf
var testFetcher = &test.Fetcher{}
var blobA = &test.Blob{"AAAAAaaaaa"}

View File

@ -30,8 +30,6 @@ import (
"camlistore.org/pkg/singleflight"
)
var _ = log.Printf
const closedIndex = -1
var errClosed = errors.New("filereader is closed")

View File

@ -21,7 +21,6 @@ import (
"bytes"
"fmt"
"io"
"log"
"os"
"strings"
@ -59,8 +58,6 @@ const (
tooSmallThreshold = 64 << 10
)
var _ = log.Printf
// WriteFileFromReader creates and uploads a "file" JSON schema
// composed of chunks of r, also uploading the chunks. The returned
// BlobRef is of the JSON file schema blob.

View File

@ -35,8 +35,6 @@ var queueSyncInterval = 5 * time.Second
const maxErrors = 20
var _ = log.Printf
// TODO: rate control + tunable
// TODO: expose copierPoolSize as tunable
type SyncHandler struct {

View File

@ -23,7 +23,6 @@ import (
"fmt"
"io"
"io/ioutil"
"log"
"net/http"
"os"
"regexp"
@ -40,8 +39,6 @@ import (
"camlistore.org/pkg/jsonconfig"
)
var _ = log.Printf
const (
blobKind = "Blob"
memKind = "NsBlobMember" // blob membership in a namespace