2011-01-28 07:07:18 +00:00
|
|
|
/*
|
|
|
|
Copyright 2011 Google Inc.
|
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License.
|
|
|
|
*/
|
|
|
|
|
2010-12-30 18:17:47 +00:00
|
|
|
package schema
|
|
|
|
|
|
|
|
import (
|
2011-01-01 10:51:15 +00:00
|
|
|
"bufio"
|
2010-12-30 18:17:47 +00:00
|
|
|
"bytes"
|
2010-12-31 06:37:46 +00:00
|
|
|
"camli/blobref"
|
|
|
|
"crypto/sha1"
|
2010-12-30 18:17:47 +00:00
|
|
|
"fmt"
|
2010-12-31 06:37:46 +00:00
|
|
|
"io"
|
2010-12-30 18:17:47 +00:00
|
|
|
"json"
|
|
|
|
"os"
|
2011-01-16 02:10:58 +00:00
|
|
|
"rand"
|
2011-01-01 10:51:15 +00:00
|
|
|
"strconv"
|
2010-12-30 18:17:47 +00:00
|
|
|
"strings"
|
2011-01-01 10:51:15 +00:00
|
|
|
"sync"
|
2011-03-23 05:59:45 +00:00
|
|
|
"syscall"
|
2010-12-31 06:37:46 +00:00
|
|
|
"time"
|
2010-12-30 18:17:47 +00:00
|
|
|
)
|
|
|
|
|
2010-12-31 06:37:46 +00:00
|
|
|
var NoCamliVersionError = os.NewError("No camliVersion key in map")
|
|
|
|
var UnimplementedError = os.NewError("Unimplemented")
|
|
|
|
|
|
|
|
type StatHasher interface {
|
|
|
|
Lstat(fileName string) (*os.FileInfo, os.Error)
|
|
|
|
Hash(fileName string) (*blobref.BlobRef, os.Error)
|
2010-12-30 18:17:47 +00:00
|
|
|
}
|
|
|
|
|
2011-03-13 18:38:53 +00:00
|
|
|
// Superset represents the superset of common camlistore JSON schema
|
|
|
|
// keys as a convenient json.Unmarshal target
|
|
|
|
type Superset struct {
|
2011-03-24 05:04:50 +00:00
|
|
|
BlobRef *blobref.BlobRef // Not in JSON, but included for
|
|
|
|
// those who want to set it.
|
|
|
|
|
2011-03-13 18:38:53 +00:00
|
|
|
Version int "camliVersion"
|
|
|
|
Type string "camliType"
|
|
|
|
|
|
|
|
Signer string "camliSigner"
|
|
|
|
Sig string "camliSig"
|
|
|
|
|
|
|
|
ClaimType string "claimType"
|
|
|
|
ClaimDate string "claimDate"
|
|
|
|
|
|
|
|
Permanode string "permaNode"
|
|
|
|
Attribute string "attribute"
|
|
|
|
Value string "value"
|
2011-03-23 05:11:27 +00:00
|
|
|
|
|
|
|
FileName string "fileName"
|
|
|
|
FileNameBytes []byte "fileNameBytes" // TODO: needs custom UnmarshalJSON?
|
|
|
|
UnixPermission string "unixPermission"
|
|
|
|
UnixOwnerId int "unixOwnerId"
|
|
|
|
UnixOwner string "unixOwner"
|
|
|
|
UnixGroupId int "unixGroupId"
|
|
|
|
UnixGroup string "unixGroup"
|
|
|
|
UnixMtime string "unixMtime"
|
|
|
|
UnixCtime string "unixCtime"
|
|
|
|
UnixAtime string "unixAtime"
|
|
|
|
|
2011-03-24 22:33:15 +00:00
|
|
|
Size uint64 "size" // for files
|
2011-03-24 05:04:50 +00:00
|
|
|
ContentParts []*ContentPart "contentParts"
|
2011-03-23 05:32:11 +00:00
|
|
|
|
2011-03-23 05:11:27 +00:00
|
|
|
Entries string "entries" // for directories, a blobref to a static-set
|
|
|
|
Members []string "members" // for static sets (for directory static-sets:
|
|
|
|
// blobrefs to child dirs/files)
|
|
|
|
}
|
|
|
|
|
2011-03-24 05:04:50 +00:00
|
|
|
type ContentPart struct {
|
|
|
|
BlobRefString string "blobRef"
|
|
|
|
BlobRef *blobref.BlobRef // TODO: ditch BlobRefString? use json.Unmarshaler?
|
2011-03-24 22:33:15 +00:00
|
|
|
Size uint64 "size"
|
|
|
|
Offset uint64 "offset"
|
|
|
|
}
|
|
|
|
|
|
|
|
func (cp *ContentPart) blobref() *blobref.BlobRef {
|
|
|
|
if cp.BlobRef == nil {
|
|
|
|
cp.BlobRef = blobref.Parse(cp.BlobRefString)
|
|
|
|
}
|
|
|
|
return cp.BlobRef
|
2011-03-24 05:04:50 +00:00
|
|
|
}
|
|
|
|
|
2011-03-23 05:11:27 +00:00
|
|
|
func (ss *Superset) HasFilename(name string) bool {
|
|
|
|
// TODO: use filename bytes too
|
|
|
|
return ss.FileName == name
|
2011-03-13 18:38:53 +00:00
|
|
|
}
|
|
|
|
|
2011-03-23 05:59:45 +00:00
|
|
|
func (ss *Superset) UnixMode() (mode uint32) {
|
|
|
|
m64, err := strconv.Btoui64(ss.UnixPermission, 8)
|
|
|
|
if err == nil {
|
|
|
|
mode = mode | uint32(m64)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: add other types
|
|
|
|
switch ss.Type {
|
|
|
|
case "directory":
|
|
|
|
mode = mode | syscall.S_IFDIR
|
|
|
|
case "file":
|
|
|
|
mode = mode | syscall.S_IFREG
|
|
|
|
case "symlink":
|
|
|
|
mode = mode | syscall.S_IFLNK
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2011-03-24 22:33:15 +00:00
|
|
|
type FileReader struct {
|
|
|
|
fetcher blobref.Fetcher
|
|
|
|
ss *Superset
|
|
|
|
ci int // index into contentparts
|
|
|
|
ccon uint64 // bytes into current chunk already consumed
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ss *Superset) NewFileReader(fetcher blobref.Fetcher) *FileReader {
|
|
|
|
return &FileReader{fetcher, ss, 0, 0}
|
|
|
|
}
|
|
|
|
|
|
|
|
func minu64(a, b uint64) uint64 {
|
|
|
|
if a < b {
|
|
|
|
return a
|
|
|
|
}
|
|
|
|
return b
|
|
|
|
}
|
|
|
|
|
|
|
|
func (fr *FileReader) Skip(skipBytes uint64) {
|
|
|
|
for skipBytes != 0 && fr.ci < len(fr.ss.ContentParts) {
|
|
|
|
cp := fr.ss.ContentParts[fr.ci]
|
|
|
|
thisChunkSkippable := cp.Size - fr.ccon
|
|
|
|
toSkip := minu64(skipBytes, thisChunkSkippable)
|
|
|
|
fr.ccon += toSkip
|
|
|
|
if fr.ccon == cp.Size {
|
|
|
|
fr.ci++
|
|
|
|
fr.ccon = 0
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (fr *FileReader) Read(p []byte) (n int, err os.Error) {
|
|
|
|
var cp *ContentPart
|
|
|
|
for {
|
|
|
|
if fr.ci >= len(fr.ss.ContentParts) {
|
|
|
|
return 0, os.EOF
|
|
|
|
}
|
|
|
|
cp = fr.ss.ContentParts[fr.ci]
|
|
|
|
thisChunkReadable := cp.Size - fr.ccon
|
|
|
|
if thisChunkReadable == 0 {
|
|
|
|
fr.ci++
|
|
|
|
fr.ccon = 0
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
|
|
|
|
br := cp.blobref()
|
|
|
|
if br == nil {
|
|
|
|
return 0, fmt.Errorf("no blobref in content part %d", fr.ci)
|
|
|
|
}
|
|
|
|
// TODO: performance: don't re-fetch this on every
|
|
|
|
// Read call. most parts will be large relative to
|
|
|
|
// read sizes. we should stuff the rsc away in fr
|
|
|
|
// and re-use it just re-seeking if needed, which
|
|
|
|
// could also be tracked.
|
|
|
|
rsc, _, ferr := fr.fetcher.Fetch(br)
|
|
|
|
if ferr != nil {
|
|
|
|
return 0, fmt.Errorf("schema: FileReader.Read error fetching blob %s: %v", br, ferr)
|
|
|
|
}
|
|
|
|
defer rsc.Close()
|
|
|
|
|
|
|
|
seekTo := cp.Offset + fr.ccon
|
|
|
|
if seekTo != 0 {
|
|
|
|
_, serr := rsc.Seek(int64(seekTo), 0)
|
|
|
|
if serr != nil {
|
|
|
|
return 0, fmt.Errorf("schema: FileReader.Read seek error on blob %s: %v", br, serr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return rsc.Read(p)
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2010-12-31 06:37:46 +00:00
|
|
|
var DefaultStatHasher = &defaultStatHasher{}
|
|
|
|
|
|
|
|
type defaultStatHasher struct{}
|
|
|
|
|
|
|
|
func (d *defaultStatHasher) Lstat(fileName string) (*os.FileInfo, os.Error) {
|
|
|
|
return os.Lstat(fileName)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *defaultStatHasher) Hash(fileName string) (*blobref.BlobRef, os.Error) {
|
|
|
|
s1 := sha1.New()
|
|
|
|
file, err := os.Open(fileName, os.O_RDONLY, 0)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
defer file.Close()
|
|
|
|
_, err = io.Copy(s1, file)
|
2011-03-13 18:39:08 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2010-12-31 06:37:46 +00:00
|
|
|
return blobref.FromHash("sha1", s1), nil
|
|
|
|
}
|
2010-12-30 18:17:47 +00:00
|
|
|
|
2011-01-01 23:44:08 +00:00
|
|
|
type StaticSet struct {
|
2011-03-13 18:39:08 +00:00
|
|
|
l sync.Mutex
|
2011-01-01 23:44:08 +00:00
|
|
|
refs []*blobref.BlobRef
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ss *StaticSet) Add(ref *blobref.BlobRef) {
|
|
|
|
ss.l.Lock()
|
|
|
|
defer ss.l.Unlock()
|
|
|
|
ss.refs = append(ss.refs, ref)
|
|
|
|
}
|
|
|
|
|
|
|
|
func newCamliMap(version int, ctype string) map[string]interface{} {
|
|
|
|
m := make(map[string]interface{})
|
2011-03-13 18:39:08 +00:00
|
|
|
m["camliVersion"] = version
|
|
|
|
m["camliType"] = ctype
|
2011-01-01 23:44:08 +00:00
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
2011-01-17 05:16:54 +00:00
|
|
|
func NewUnsignedPermanode() map[string]interface{} {
|
2011-01-16 02:10:58 +00:00
|
|
|
m := newCamliMap(1, "permanode")
|
|
|
|
chars := make([]byte, 20)
|
|
|
|
// Don't need cryptographically secure random here, as this
|
|
|
|
// will be GPG signed anyway.
|
2011-01-16 04:20:35 +00:00
|
|
|
rnd := rand.New(rand.NewSource(time.Nanoseconds()))
|
2011-01-16 02:10:58 +00:00
|
|
|
for idx, _ := range chars {
|
2011-03-13 18:39:08 +00:00
|
|
|
chars[idx] = byte(32 + rnd.Intn(126-32))
|
2011-01-16 02:10:58 +00:00
|
|
|
}
|
|
|
|
m["random"] = string(chars)
|
2011-01-17 05:16:54 +00:00
|
|
|
return m
|
2011-01-16 02:10:58 +00:00
|
|
|
}
|
|
|
|
|
2011-01-01 23:44:08 +00:00
|
|
|
// Map returns a Camli map of camliType "static-set"
|
|
|
|
func (ss *StaticSet) Map() map[string]interface{} {
|
|
|
|
m := newCamliMap(1, "static-set")
|
|
|
|
ss.l.Lock()
|
|
|
|
defer ss.l.Unlock()
|
|
|
|
|
|
|
|
members := make([]string, 0, len(ss.refs))
|
|
|
|
if ss.refs != nil {
|
|
|
|
for _, ref := range ss.refs {
|
|
|
|
members = append(members, ref.String())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
m["members"] = members
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
2010-12-30 18:17:47 +00:00
|
|
|
func MapToCamliJson(m map[string]interface{}) (string, os.Error) {
|
|
|
|
version, hasVersion := m["camliVersion"]
|
|
|
|
if !hasVersion {
|
2010-12-31 06:37:46 +00:00
|
|
|
return "", NoCamliVersionError
|
2010-12-30 18:17:47 +00:00
|
|
|
}
|
|
|
|
m["camliVersion"] = 0, false
|
|
|
|
jsonBytes, err := json.MarshalIndent(m, "", " ")
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
m["camliVersion"] = version
|
|
|
|
buf := new(bytes.Buffer)
|
|
|
|
fmt.Fprintf(buf, "{\"camliVersion\": %v,\n", version)
|
|
|
|
buf.Write(jsonBytes[2:])
|
|
|
|
return string(buf.Bytes()), nil
|
|
|
|
}
|
|
|
|
|
2011-01-01 22:44:19 +00:00
|
|
|
func NewCommonFileMap(fileName string, fi *os.FileInfo) map[string]interface{} {
|
2011-03-13 18:39:08 +00:00
|
|
|
m := newCamliMap(1, "" /* no type yet */ )
|
|
|
|
|
2010-12-30 18:17:47 +00:00
|
|
|
lastSlash := strings.LastIndex(fileName, "/")
|
|
|
|
baseName := fileName[lastSlash+1:]
|
|
|
|
if isValidUtf8(baseName) {
|
2010-12-31 06:37:46 +00:00
|
|
|
m["fileName"] = baseName
|
2010-12-30 18:17:47 +00:00
|
|
|
} else {
|
2010-12-31 06:37:46 +00:00
|
|
|
m["fileNameBytes"] = []uint8(baseName)
|
|
|
|
}
|
2011-01-01 22:44:19 +00:00
|
|
|
|
|
|
|
// Common elements (from file-common.txt)
|
2011-02-03 16:19:40 +00:00
|
|
|
if !fi.IsSymlink() {
|
|
|
|
m["unixPermission"] = fmt.Sprintf("0%o", fi.Permission())
|
|
|
|
}
|
2011-01-01 22:44:19 +00:00
|
|
|
if fi.Uid != -1 {
|
|
|
|
m["unixOwnerId"] = fi.Uid
|
|
|
|
if user := getUserFromUid(fi.Uid); user != "" {
|
|
|
|
m["unixOwner"] = user
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if fi.Gid != -1 {
|
|
|
|
m["unixGroupId"] = fi.Gid
|
|
|
|
if group := getGroupFromGid(fi.Gid); group != "" {
|
|
|
|
m["unixGroup"] = group
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if mtime := fi.Mtime_ns; mtime != 0 {
|
|
|
|
m["unixMtime"] = rfc3339FromNanos(mtime)
|
|
|
|
}
|
2011-01-02 03:37:39 +00:00
|
|
|
// Include the ctime too, if it differs.
|
|
|
|
if ctime := fi.Ctime_ns; ctime != 0 && fi.Mtime_ns != fi.Ctime_ns {
|
|
|
|
m["unixCtime"] = rfc3339FromNanos(ctime)
|
|
|
|
}
|
2011-01-01 22:44:19 +00:00
|
|
|
|
2010-12-31 06:37:46 +00:00
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
2011-01-01 22:44:19 +00:00
|
|
|
type InvalidContentPartsError struct {
|
|
|
|
StatSize int64
|
|
|
|
SumOfParts int64
|
|
|
|
}
|
|
|
|
|
|
|
|
func (e *InvalidContentPartsError) String() string {
|
|
|
|
return fmt.Sprintf("Invalid ContentPart slice in PopulateRegularFileMap; file stat size is %d but sum of parts was %d", e.StatSize, e.SumOfParts)
|
|
|
|
}
|
|
|
|
|
|
|
|
func PopulateRegularFileMap(m map[string]interface{}, fi *os.FileInfo, parts []ContentPart) os.Error {
|
2010-12-31 06:37:46 +00:00
|
|
|
m["camliType"] = "file"
|
|
|
|
m["size"] = fi.Size
|
|
|
|
|
2011-03-24 22:33:15 +00:00
|
|
|
sumSize := uint64(0)
|
2011-01-01 22:44:19 +00:00
|
|
|
mparts := make([]map[string]interface{}, len(parts))
|
|
|
|
for idx, part := range parts {
|
|
|
|
mpart := make(map[string]interface{})
|
|
|
|
mparts[idx] = mpart
|
|
|
|
mpart["blobRef"] = part.BlobRef.String()
|
|
|
|
mpart["size"] = part.Size
|
|
|
|
sumSize += part.Size
|
|
|
|
if part.Offset != 0 {
|
|
|
|
mpart["offset"] = part.Offset
|
|
|
|
}
|
|
|
|
}
|
2011-03-24 22:33:15 +00:00
|
|
|
if sumSize != uint64(fi.Size) {
|
|
|
|
return &InvalidContentPartsError{fi.Size, int64(sumSize)}
|
2010-12-30 18:17:47 +00:00
|
|
|
}
|
2011-01-01 22:44:19 +00:00
|
|
|
m["contentParts"] = mparts
|
2011-01-01 21:12:45 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2011-01-01 22:44:19 +00:00
|
|
|
func PopulateSymlinkMap(m map[string]interface{}, fileName string) os.Error {
|
2011-01-01 21:12:45 +00:00
|
|
|
m["camliType"] = "symlink"
|
|
|
|
target, err := os.Readlink(fileName)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if isValidUtf8(target) {
|
|
|
|
m["symlinkTarget"] = target
|
|
|
|
} else {
|
|
|
|
m["symlinkTargetBytes"] = []uint8(target)
|
|
|
|
}
|
2010-12-31 06:37:46 +00:00
|
|
|
return nil
|
2010-12-30 18:17:47 +00:00
|
|
|
}
|
|
|
|
|
2011-01-01 23:44:08 +00:00
|
|
|
func PopulateDirectoryMap(m map[string]interface{}, staticSetRef *blobref.BlobRef) {
|
|
|
|
m["camliType"] = "directory"
|
|
|
|
m["entries"] = staticSetRef.String()
|
|
|
|
}
|
|
|
|
|
2011-01-26 06:44:03 +00:00
|
|
|
func NewShareRef(authType string, target *blobref.BlobRef, transitive bool) map[string]interface{} {
|
2011-03-13 03:28:18 +00:00
|
|
|
m := newCamliMap(1, "share")
|
2011-01-26 06:44:03 +00:00
|
|
|
m["authType"] = authType
|
|
|
|
m["target"] = target.String()
|
|
|
|
m["transitive"] = transitive
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
2011-03-13 03:28:18 +00:00
|
|
|
func NewClaim(permaNode *blobref.BlobRef, claimType string) map[string]interface{} {
|
|
|
|
m := newCamliMap(1, "claim")
|
|
|
|
m["permaNode"] = permaNode.String()
|
|
|
|
m["claimType"] = claimType
|
2011-03-13 19:30:13 +00:00
|
|
|
m["claimDate"] = rfc3339FromNanos(time.Nanoseconds())
|
2011-03-13 03:28:18 +00:00
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
func newAttrChangeClaim(permaNode *blobref.BlobRef, claimType, attr, value string) map[string]interface{} {
|
2011-03-14 05:21:58 +00:00
|
|
|
m := NewClaim(permaNode, claimType)
|
2011-03-13 03:28:18 +00:00
|
|
|
m["attribute"] = attr
|
|
|
|
m["value"] = value
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewSetAttributeClaim(permaNode *blobref.BlobRef, attr, value string) map[string]interface{} {
|
|
|
|
return newAttrChangeClaim(permaNode, "set-attribute", attr, value)
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewAddAttributeClaim(permaNode *blobref.BlobRef, attr, value string) map[string]interface{} {
|
|
|
|
return newAttrChangeClaim(permaNode, "add-attribute", attr, value)
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewDelAttributeClaim(permaNode *blobref.BlobRef, attr string) map[string]interface{} {
|
|
|
|
m := newAttrChangeClaim(permaNode, "del-attribute", attr, "")
|
|
|
|
m["value"] = "", false
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
2011-01-26 06:44:03 +00:00
|
|
|
// Types of ShareRefs
|
|
|
|
const ShareHaveRef = "haveref"
|
|
|
|
|
2010-12-31 20:13:33 +00:00
|
|
|
func rfc3339FromNanos(epochnanos int64) string {
|
|
|
|
nanos := epochnanos % 1e9
|
|
|
|
esec := epochnanos / 1e9
|
|
|
|
t := time.SecondsToUTC(esec)
|
|
|
|
timeStr := t.Format(time.RFC3339)
|
|
|
|
if nanos == 0 {
|
|
|
|
return timeStr
|
|
|
|
}
|
|
|
|
nanoStr := fmt.Sprintf("%09d", nanos)
|
|
|
|
nanoStr = strings.TrimRight(nanoStr, "0")
|
|
|
|
return timeStr[:len(timeStr)-1] + "." + nanoStr + "Z"
|
|
|
|
}
|
|
|
|
|
2011-01-01 10:51:15 +00:00
|
|
|
func populateMap(m map[int]string, file string) {
|
|
|
|
f, err := os.Open(file, os.O_RDONLY, 0)
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
bufr := bufio.NewReader(f)
|
|
|
|
for {
|
|
|
|
line, err := bufr.ReadString('\n')
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
parts := strings.Split(line, ":", 4)
|
|
|
|
if len(parts) >= 3 {
|
|
|
|
idstr := parts[2]
|
|
|
|
id, err := strconv.Atoi(idstr)
|
|
|
|
if err == nil {
|
|
|
|
m[id] = parts[0]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var uidToUsernameMap map[int]string
|
|
|
|
var getUserFromUidOnce sync.Once
|
2011-03-13 18:39:08 +00:00
|
|
|
|
2011-01-01 10:51:15 +00:00
|
|
|
func getUserFromUid(uid int) string {
|
|
|
|
getUserFromUidOnce.Do(func() {
|
|
|
|
uidToUsernameMap = make(map[int]string)
|
|
|
|
populateMap(uidToUsernameMap, "/etc/passwd")
|
|
|
|
})
|
|
|
|
return uidToUsernameMap[uid]
|
|
|
|
}
|
|
|
|
|
|
|
|
var gidToUsernameMap map[int]string
|
|
|
|
var getGroupFromGidOnce sync.Once
|
2011-03-13 18:39:08 +00:00
|
|
|
|
2011-01-01 10:51:15 +00:00
|
|
|
func getGroupFromGid(uid int) string {
|
|
|
|
getGroupFromGidOnce.Do(func() {
|
|
|
|
gidToUsernameMap = make(map[int]string)
|
|
|
|
populateMap(gidToUsernameMap, "/etc/group")
|
|
|
|
})
|
|
|
|
return gidToUsernameMap[uid]
|
|
|
|
}
|
|
|
|
|
2010-12-31 06:37:46 +00:00
|
|
|
func isValidUtf8(s string) bool {
|
|
|
|
for _, rune := range []int(s) {
|
|
|
|
if rune == 0xfffd {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true
|
2010-12-30 18:17:47 +00:00
|
|
|
}
|