diff --git a/pytorch3d/implicitron/models/renderer/sdf_renderer.py b/pytorch3d/implicitron/models/renderer/sdf_renderer.py index 735a8a8c..e06f42d3 100644 --- a/pytorch3d/implicitron/models/renderer/sdf_renderer.py +++ b/pytorch3d/implicitron/models/renderer/sdf_renderer.py @@ -219,7 +219,7 @@ def _sample_network( surface_dists, surface_cam_loc, surface_ray_dirs, - eps=1e-4, + eps: float = 1e-4, ): # t -> t(theta) surface_ray_dirs_0 = surface_ray_dirs.detach() diff --git a/pytorch3d/implicitron/models/view_pooling/feature_aggregation.py b/pytorch3d/implicitron/models/view_pooling/feature_aggregation.py index 01022607..1ed6b481 100644 --- a/pytorch3d/implicitron/models/view_pooling/feature_aggregation.py +++ b/pytorch3d/implicitron/models/view_pooling/feature_aggregation.py @@ -412,7 +412,7 @@ class AngleWeightedIdentityFeatureAggregator(torch.nn.Module, FeatureAggregatorB def _get_reduction_aggregator_feature_dim( feats_or_feats_dim: Union[Dict[str, torch.Tensor], int], reduction_functions: Sequence[ReductionFunction], -): +) -> int: if isinstance(feats_or_feats_dim, int): feat_dim = feats_or_feats_dim else: diff --git a/pytorch3d/ops/points_normals.py b/pytorch3d/ops/points_normals.py index 70128814..f9113299 100644 --- a/pytorch3d/ops/points_normals.py +++ b/pytorch3d/ops/points_normals.py @@ -167,7 +167,7 @@ def estimate_pointcloud_local_coord_frames( return curvatures, local_coord_frames -def _disambiguate_vector_directions(pcl, knns, vecs): +def _disambiguate_vector_directions(pcl, knns, vecs: float) -> float: """ Disambiguates normal directions according to [1]. @@ -181,6 +181,7 @@ def _disambiguate_vector_directions(pcl, knns, vecs): # each element of the neighborhood df = knns - pcl[:, :, None] # projection of the difference on the principal direction + # pyre-fixme[16]: `float` has no attribute `__getitem__`. proj = (vecs[:, :, None] * df).sum(3) # check how many projections are positive n_pos = (proj > 0).type_as(knns).sum(2, keepdim=True) diff --git a/pytorch3d/ops/subdivide_meshes.py b/pytorch3d/ops/subdivide_meshes.py index dec3be74..2dd0cab4 100644 --- a/pytorch3d/ops/subdivide_meshes.py +++ b/pytorch3d/ops/subdivide_meshes.py @@ -400,7 +400,7 @@ def create_verts_index(verts_per_mesh, edges_per_mesh, device=None): return verts_idx -def create_faces_index(faces_per_mesh, device=None): +def create_faces_index(faces_per_mesh: int, device=None): """ Helper function to group the faces indices for each mesh. New faces are stacked at the end of the original faces tensor, so in order to have @@ -417,7 +417,9 @@ def create_faces_index(faces_per_mesh, device=None): """ # e.g. faces_per_mesh = [2, 5, 3] + # pyre-fixme[16]: `int` has no attribute `sum`. F = faces_per_mesh.sum() # e.g. 10 + # pyre-fixme[16]: `int` has no attribute `cumsum`. faces_per_mesh_cumsum = faces_per_mesh.cumsum(dim=0) # (N,) e.g. (2, 7, 10) switch1_idx = faces_per_mesh_cumsum.clone() diff --git a/pytorch3d/renderer/cameras.py b/pytorch3d/renderer/cameras.py index 42fd6471..3b660c3c 100644 --- a/pytorch3d/renderer/cameras.py +++ b/pytorch3d/renderer/cameras.py @@ -427,10 +427,10 @@ class CamerasBase(TensorProperties): def OpenGLPerspectiveCameras( - znear=1.0, - zfar=100.0, - aspect_ratio=1.0, - fov=60.0, + znear: float = 1.0, + zfar: float = 100.0, + aspect_ratio: float = 1.0, + fov: float = 60.0, degrees: bool = True, R: torch.Tensor = _R, T: torch.Tensor = _T, @@ -709,12 +709,12 @@ class FoVPerspectiveCameras(CamerasBase): def OpenGLOrthographicCameras( - znear=1.0, - zfar=100.0, - top=1.0, - bottom=-1.0, - left=-1.0, - right=1.0, + znear: float = 1.0, + zfar: float = 100.0, + top: float = 1.0, + bottom: float = -1.0, + left: float = -1.0, + right: float = 1.0, scale_xyz=((1.0, 1.0, 1.0),), # (1, 3) R: torch.Tensor = _R, T: torch.Tensor = _T, @@ -956,7 +956,7 @@ Note that the MultiView Cameras accept parameters in NDC space. def SfMPerspectiveCameras( - focal_length=1.0, + focal_length: float = 1.0, principal_point=((0.0, 0.0),), R: torch.Tensor = _R, T: torch.Tensor = _T, @@ -1194,7 +1194,7 @@ class PerspectiveCameras(CamerasBase): def SfMOrthographicCameras( - focal_length=1.0, + focal_length: float = 1.0, principal_point=((0.0, 0.0),), R: torch.Tensor = _R, T: torch.Tensor = _T, @@ -1645,9 +1645,9 @@ def look_at_rotation( def look_at_view_transform( - dist=1.0, - elev=0.0, - azim=0.0, + dist: float = 1.0, + elev: float = 0.0, + azim: float = 0.0, degrees: bool = True, eye: Optional[Union[Sequence, torch.Tensor]] = None, at=((0, 0, 0),), # (1, 3) diff --git a/pytorch3d/vis/texture_vis.py b/pytorch3d/vis/texture_vis.py index ee920c38..a3b0fcd5 100644 --- a/pytorch3d/vis/texture_vis.py +++ b/pytorch3d/vis/texture_vis.py @@ -65,7 +65,7 @@ def texturesuv_image_PIL( *, texture_index: int = 0, radius: float = 1, - color="red", + color: str = "red", subsample: Optional[int] = 10000, ): # pragma: no cover """